use_pty:FALSE /usr/share/restraint/plugins/run_task_plugins bash ./runtest.sh :::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::: :: Test :::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::: :: [ 00:18:36 ] :: [ BEGIN ] :: Running 'uname -a' Linux hpe-dl360pgen8-08.hpe2.lab.eng.bos.redhat.com 5.14.0-247.rt14.248.1955_761038739.el9.x86_64+debug #1 SMP PREEMPT_RT Mon Jan 30 00:40:38 UTC 2023 x86_64 x86_64 x86_64 GNU/Linux :: [ 00:18:36 ] :: [ PASS ] :: Command 'uname -a' (Expected 0, got 0) :: [ 00:18:38 ] :: [ BEGIN ] :: Running 'rpm -q mdadm || dnf install -y mdadm' mdadm-4.2-8.el9.x86_64 :: [ 00:18:38 ] :: [ PASS ] :: Command 'rpm -q mdadm || dnf install -y mdadm' (Expected 0, got 0) :: [ 00:18:38 ] :: [ LOG ] :: ./runtest.sh :: [ 00:18:38 ] :: [ BEGIN ] :: Running 'modprobe raid456 devices_handle_discard_safely=Y' :: [ 00:18:39 ] :: [ PASS ] :: Command 'modprobe raid456 devices_handle_discard_safely=Y' (Expected 0, got 0) :: [ 00:18:39 ] :: [ BEGIN ] :: Running 'echo Y >/sys/module/raid456/parameters/devices_handle_discard_safely' :: [ 00:18:39 ] :: [ PASS ] :: Command 'echo Y >/sys/module/raid456/parameters/devices_handle_discard_safely' (Expected 0, got 0) /usr/sbin/mkfs.xfs INFO: Executing MD_Create_RAID() to create raid 0 INFO: Created md raid with these raid devices " /dev/loop0 /dev/loop3 /dev/loop1 /dev/loop2 /dev/loop4". :: [ 00:19:07 ] :: [ BEGIN ] :: Running 'mdadm --create --run /dev/md1 --level 0 --metadata 1.2 --raid-devices 5 /dev/loop0 /dev/loop3 /dev/loop1 /dev/loop2 /dev/loop4 --chunk 512' mdadm: array /dev/md1 started. :: [ 00:19:35 ] :: [ PASS ] :: Command 'mdadm --create --run /dev/md1 --level 0 --metadata 1.2 --raid-devices 5 /dev/loop0 /dev/loop3 /dev/loop1 /dev/loop2 /dev/loop4 --chunk 512' (Expected 0, got 0) INFO:cat /proc/mdstat###################### :: [ 00:19:36 ] :: [ BEGIN ] :: Running 'cat /proc/mdstat' Personalities : [raid6] [raid5] [raid4] [raid0] md1 : active raid0 loop4[4] loop2[3] loop1[2] loop3[1] loop0[0] 2549760 blocks super 1.2 512k chunks unused devices: :: [ 00:19:36 ] :: [ PASS ] :: Command 'cat /proc/mdstat' (Expected 0, got 0) :: [ 00:19:36 ] :: [ BEGIN ] :: Running 'lsblk' NAME MAJ:MIN RM SIZE RO TYPE MOUNTPOINTS loop0 7:0 0 500M 0 loop └─md1 9:1 0 2.4G 0 raid0 loop1 7:1 0 500M 0 loop └─md1 9:1 0 2.4G 0 raid0 loop2 7:2 0 500M 0 loop └─md1 9:1 0 2.4G 0 raid0 loop3 7:3 0 500M 0 loop └─md1 9:1 0 2.4G 0 raid0 loop4 7:4 0 500M 0 loop └─md1 9:1 0 2.4G 0 raid0 loop5 7:5 0 500M 0 loop sda 8:0 0 465.7G 0 disk ├─sda1 8:1 0 1G 0 part /boot └─sda2 8:2 0 464.7G 0 part ├─cs_hpe--dl360pgen8--08-root 253:0 0 70G 0 lvm / ├─cs_hpe--dl360pgen8--08-swap 253:1 0 15.7G 0 lvm [SWAP] └─cs_hpe--dl360pgen8--08-home 253:2 0 379G 0 lvm /home :: [ 00:19:36 ] :: [ PASS ] :: Command 'lsblk' (Expected 0, got 0) /dev/md1 INFO:mdadm -D /dev/md1 ######################### :: [ 00:19:37 ] :: [ BEGIN ] :: Running 'mdadm --detail /dev/md1' /dev/md1: Version : 1.2 Creation Time : Mon Jan 30 00:19:07 2023 Raid Level : raid0 Array Size : 2549760 (2.43 GiB 2.61 GB) Raid Devices : 5 Total Devices : 5 Persistence : Superblock is persistent Update Time : Mon Jan 30 00:19:07 2023 State : clean Active Devices : 5 Working Devices : 5 Failed Devices : 0 Spare Devices : 0 Layout : -unknown- Chunk Size : 512K Consistency Policy : none Name : 1 UUID : 1408ad61:3bfd6f93:630f2595:1d36cd8b Events : 0 Number Major Minor RaidDevice State 0 7 0 0 active sync /dev/loop0 1 7 3 1 active sync /dev/loop3 2 7 1 2 active sync /dev/loop1 3 7 2 3 active sync /dev/loop2 4 7 4 4 active sync /dev/loop4 :: [ 00:19:37 ] :: [ PASS ] :: Command 'mdadm --detail /dev/md1' (Expected 0, got 0) :: [ 00:19:37 ] :: [ LOG ] :: INFO: Successfully created md raid /dev/md1 1675055977 start_time against this md array: /dev/md1 state is clean :: [ 00:19:38 ] :: [ LOG ] :: mkfs -t xfs /dev/md1 log stripe unit (524288 bytes) is too large (maximum is 256KiB) log stripe unit adjusted to 32KiB meta-data=/dev/md1 isize=512 agcount=8, agsize=79744 blks = sectsz=512 attr=2, projid32bit=1 = crc=1 finobt=1, sparse=1, rmapbt=0 = reflink=1 bigtime=1 inobtcount=1 data = bsize=4096 blocks=637440, imaxpct=25 = sunit=128 swidth=640 blks naming =version 2 bsize=4096 ascii-ci=0, ftype=1 log =internal log bsize=4096 blocks=2560, version=2 = sectsz=512 sunit=8 blks, lazy-count=1 realtime =none extsz=4096 blocks=0, rtextents=0 Discarding blocks...Done. :: [ 00:19:40 ] :: [ BEGIN ] :: Running 'mount -t xfs /dev/md1 /mnt/md_test ' :: [ 00:19:40 ] :: [ PASS ] :: Command 'mount -t xfs /dev/md1 /mnt/md_test ' (Expected 0, got 0) :: [ 00:19:40 ] :: [ BEGIN ] :: Running 'fstrim -v /mnt/md_test' /mnt/md_test: 2.4 GiB (2600108032 bytes) trimmed :: [ 00:19:41 ] :: [ PASS ] :: Command 'fstrim -v /mnt/md_test' (Expected 0, got 0) :: [ 00:19:41 ] :: [ BEGIN ] :: Running 'umount /dev/md1' :: [ 00:19:42 ] :: [ PASS ] :: Command 'umount /dev/md1' (Expected 0, got 0) INFO: Executing MD_Clean_RAID() against this md device: mdadm --stop /dev/md1 mdadm: stopped /dev/md1 clean devs : /dev/loop0 /dev/loop3 /dev/loop1 /dev/loop2 /dev/loop4 mdadm --zero-superblock /dev/loop0 mdadm --zero-superblock /dev/loop3 mdadm --zero-superblock /dev/loop1 mdadm --zero-superblock /dev/loop2 mdadm --zero-superblock /dev/loop4 ret is 0 ls /dev/md1 ls: cannot access '/dev/md1': No such file or directory mdadm --stop can delete md node name /dev/md1 in /dev INFO: Executing MD_Create_RAID() to create raid 1 INFO: Created md raid with these raid devices " /dev/loop0 /dev/loop3 /dev/loop1 /dev/loop2 /dev/loop4". INFO: Created md raid with these spare disks " /dev/loop5". :: [ 00:20:09 ] :: [ BEGIN ] :: Running 'mdadm --create --run /dev/md1 --level 1 --metadata 1.2 --raid-devices 5 /dev/loop0 /dev/loop3 /dev/loop1 /dev/loop2 /dev/loop4 --spare-devices 1 /dev/loop5 --bitmap=internal --bitmap-chunk=64M' mdadm: array /dev/md1 started. :: [ 00:20:09 ] :: [ PASS ] :: Command 'mdadm --create --run /dev/md1 --level 1 --metadata 1.2 --raid-devices 5 /dev/loop0 /dev/loop3 /dev/loop1 /dev/loop2 /dev/loop4 --spare-devices 1 /dev/loop5 --bitmap=internal --bitmap-chunk=64M' (Expected 0, got 0) INFO:cat /proc/mdstat###################### :: [ 00:20:10 ] :: [ BEGIN ] :: Running 'cat /proc/mdstat' Personalities : [raid6] [raid5] [raid4] [raid0] [raid1] md1 : active raid1 loop5[5](S) loop4[4] loop2[3] loop1[2] loop3[1] loop0[0] 510976 blocks super 1.2 [5/5] [UUUUU] [>....................] resync = 2.3% (11776/510976) finish=0.7min speed=11776K/sec bitmap: 1/1 pages [4KB], 65536KB chunk unused devices: :: [ 00:20:10 ] :: [ PASS ] :: Command 'cat /proc/mdstat' (Expected 0, got 0) :: [ 00:20:10 ] :: [ BEGIN ] :: Running 'lsblk' NAME MAJ:MIN RM SIZE RO TYPE MOUNTPOINTS loop0 7:0 0 500M 0 loop └─md1 9:1 0 499M 0 raid1 loop1 7:1 0 500M 0 loop └─md1 9:1 0 499M 0 raid1 loop2 7:2 0 500M 0 loop └─md1 9:1 0 499M 0 raid1 loop3 7:3 0 500M 0 loop └─md1 9:1 0 499M 0 raid1 loop4 7:4 0 500M 0 loop └─md1 9:1 0 499M 0 raid1 loop5 7:5 0 500M 0 loop └─md1 9:1 0 499M 0 raid1 sda 8:0 0 465.7G 0 disk ├─sda1 8:1 0 1G 0 part /boot └─sda2 8:2 0 464.7G 0 part ├─cs_hpe--dl360pgen8--08-root 253:0 0 70G 0 lvm / ├─cs_hpe--dl360pgen8--08-swap 253:1 0 15.7G 0 lvm [SWAP] └─cs_hpe--dl360pgen8--08-home 253:2 0 379G 0 lvm /home :: [ 00:20:10 ] :: [ PASS ] :: Command 'lsblk' (Expected 0, got 0) /dev/md1 INFO:mdadm -D /dev/md1 ######################### :: [ 00:20:11 ] :: [ BEGIN ] :: Running 'mdadm --detail /dev/md1' /dev/md1: Version : 1.2 Creation Time : Mon Jan 30 00:20:09 2023 Raid Level : raid1 Array Size : 510976 (499.00 MiB 523.24 MB) Used Dev Size : 510976 (499.00 MiB 523.24 MB) Raid Devices : 5 Total Devices : 6 Persistence : Superblock is persistent Intent Bitmap : Internal Update Time : Mon Jan 30 00:20:10 2023 State : clean, resyncing Active Devices : 5 Working Devices : 6 Failed Devices : 0 Spare Devices : 1 Consistency Policy : bitmap Resync Status : 6% complete Name : 1 UUID : 065093ef:a3a502eb:06198c3e:364a3228 Events : 0 Number Major Minor RaidDevice State 0 7 0 0 active sync /dev/loop0 1 7 3 1 active sync /dev/loop3 2 7 1 2 active sync /dev/loop1 3 7 2 3 active sync /dev/loop2 4 7 4 4 active sync /dev/loop4 5 7 5 - spare /dev/loop5 :: [ 00:20:11 ] :: [ PASS ] :: Command 'mdadm --detail /dev/md1' (Expected 0, got 0) :: [ 00:20:11 ] :: [ LOG ] :: INFO: Successfully created md raid /dev/md1 1675056012 start_time against this md array: /dev/md1 state is clean, 1675056017 start_time against this md array: /dev/md1 state is clean, 1675056022 start_time against this md array: /dev/md1 state is clean, 1675056027 start_time against this md array: /dev/md1 state is clean, 1675056032 start_time against this md array: /dev/md1 state is clean, 1675056038 start_time against this md array: /dev/md1 state is clean, 1675056043 start_time against this md array: /dev/md1 state is clean, 1675056048 start_time against this md array: /dev/md1 state is clean, 1675056053 start_time against this md array: /dev/md1 state is clean, 1675056059 start_time against this md array: /dev/md1 state is clean, 1675056064 start_time against this md array: /dev/md1 state is clean, 1675056069 start_time against this md array: /dev/md1 state is clean, 1675056074 start_time against this md array: /dev/md1 state is clean :: [ 00:21:14 ] :: [ LOG ] :: mkfs -t xfs /dev/md1 meta-data=/dev/md1 isize=512 agcount=4, agsize=31936 blks = sectsz=512 attr=2, projid32bit=1 = crc=1 finobt=1, sparse=1, rmapbt=0 = reflink=1 bigtime=1 inobtcount=1 data = bsize=4096 blocks=127744, imaxpct=25 = sunit=0 swidth=0 blks naming =version 2 bsize=4096 ascii-ci=0, ftype=1 log =internal log bsize=4096 blocks=1368, version=2 = sectsz=512 sunit=0 blks, lazy-count=1 realtime =none extsz=4096 blocks=0, rtextents=0 Discarding blocks...Done. :: [ 00:21:17 ] :: [ BEGIN ] :: Running 'mount -t xfs /dev/md1 /mnt/md_test ' :: [ 00:21:17 ] :: [ PASS ] :: Command 'mount -t xfs /dev/md1 /mnt/md_test ' (Expected 0, got 0) :: [ 00:21:18 ] :: [ BEGIN ] :: Running 'fstrim -v /mnt/md_test' /mnt/md_test: 493.5 MiB (517439488 bytes) trimmed :: [ 00:21:18 ] :: [ PASS ] :: Command 'fstrim -v /mnt/md_test' (Expected 0, got 0) :: [ 00:21:18 ] :: [ BEGIN ] :: Running 'umount /dev/md1' :: [ 00:21:19 ] :: [ PASS ] :: Command 'umount /dev/md1' (Expected 0, got 0) INFO: Executing MD_Clean_RAID() against this md device: mdadm --stop /dev/md1 mdadm: stopped /dev/md1 clean devs : /dev/loop0 /dev/loop3 /dev/loop1 /dev/loop2 /dev/loop4 /dev/loop5 mdadm --zero-superblock /dev/loop0 mdadm --zero-superblock /dev/loop3 mdadm --zero-superblock /dev/loop1 mdadm --zero-superblock /dev/loop2 mdadm --zero-superblock /dev/loop4 mdadm --zero-superblock /dev/loop5 ret is 0 ls /dev/md1 ls: cannot access '/dev/md1': No such file or directory mdadm --stop can delete md node name /dev/md1 in /dev INFO: Executing MD_Create_RAID() to create raid 4 INFO: Created md raid with these raid devices " /dev/loop0 /dev/loop3 /dev/loop1 /dev/loop2 /dev/loop4". INFO: Created md raid with these spare disks " /dev/loop5". :: [ 00:21:46 ] :: [ BEGIN ] :: Running 'mdadm --create --run /dev/md1 --level 4 --metadata 1.2 --raid-devices 5 /dev/loop0 /dev/loop3 /dev/loop1 /dev/loop2 /dev/loop4 --spare-devices 1 /dev/loop5 --chunk 512 --bitmap=internal --bitmap-chunk=64M' mdadm: array /dev/md1 started. :: [ 00:21:47 ] :: [ PASS ] :: Command 'mdadm --create --run /dev/md1 --level 4 --metadata 1.2 --raid-devices 5 /dev/loop0 /dev/loop3 /dev/loop1 /dev/loop2 /dev/loop4 --spare-devices 1 /dev/loop5 --chunk 512 --bitmap=internal --bitmap-chunk=64M' (Expected 0, got 0) INFO:cat /proc/mdstat###################### :: [ 00:21:47 ] :: [ BEGIN ] :: Running 'cat /proc/mdstat' Personalities : [raid6] [raid5] [raid4] [raid0] [raid1] md1 : active raid4 loop4[6] loop5[5](S) loop2[3] loop1[2] loop3[1] loop0[0] 2039808 blocks super 1.2 level 4, 512k chunk, algorithm 0 [5/4] [UUUU_] [>....................] recovery = 0.5% (2592/509952) finish=3.2min speed=2592K/sec bitmap: 1/1 pages [4KB], 65536KB chunk unused devices: :: [ 00:21:47 ] :: [ PASS ] :: Command 'cat /proc/mdstat' (Expected 0, got 0) :: [ 00:21:47 ] :: [ BEGIN ] :: Running 'lsblk' NAME MAJ:MIN RM SIZE RO TYPE MOUNTPOINTS loop0 7:0 0 500M 0 loop └─md1 9:1 0 1.9G 0 raid4 loop1 7:1 0 500M 0 loop └─md1 9:1 0 1.9G 0 raid4 loop2 7:2 0 500M 0 loop └─md1 9:1 0 1.9G 0 raid4 loop3 7:3 0 500M 0 loop └─md1 9:1 0 1.9G 0 raid4 loop4 7:4 0 500M 0 loop └─md1 9:1 0 1.9G 0 raid4 loop5 7:5 0 500M 0 loop └─md1 9:1 0 1.9G 0 raid4 sda 8:0 0 465.7G 0 disk ├─sda1 8:1 0 1G 0 part /boot └─sda2 8:2 0 464.7G 0 part ├─cs_hpe--dl360pgen8--08-root 253:0 0 70G 0 lvm / ├─cs_hpe--dl360pgen8--08-swap 253:1 0 15.7G 0 lvm [SWAP] └─cs_hpe--dl360pgen8--08-home 253:2 0 379G 0 lvm /home :: [ 00:21:48 ] :: [ PASS ] :: Command 'lsblk' (Expected 0, got 0) /dev/md1 INFO:mdadm -D /dev/md1 ######################### :: [ 00:21:48 ] :: [ BEGIN ] :: Running 'mdadm --detail /dev/md1' /dev/md1: Version : 1.2 Creation Time : Mon Jan 30 00:21:46 2023 Raid Level : raid4 Array Size : 2039808 (1992.00 MiB 2088.76 MB) Used Dev Size : 509952 (498.00 MiB 522.19 MB) Raid Devices : 5 Total Devices : 6 Persistence : Superblock is persistent Intent Bitmap : Internal Update Time : Mon Jan 30 00:21:46 2023 State : clean, degraded, recovering Active Devices : 4 Working Devices : 6 Failed Devices : 0 Spare Devices : 2 Chunk Size : 512K Consistency Policy : bitmap Rebuild Status : 3% complete Name : 1 UUID : 5232cb80:59272284:5d011408:8343c43f Events : 1 Number Major Minor RaidDevice State 0 7 0 0 active sync /dev/loop0 1 7 3 1 active sync /dev/loop3 2 7 1 2 active sync /dev/loop1 3 7 2 3 active sync /dev/loop2 6 7 4 4 spare rebuilding /dev/loop4 5 7 5 - spare /dev/loop5 :: [ 00:21:48 ] :: [ PASS ] :: Command 'mdadm --detail /dev/md1' (Expected 0, got 0) :: [ 00:21:49 ] :: [ LOG ] :: INFO: Successfully created md raid /dev/md1 1675056109 start_time against this md array: /dev/md1 state is clean, 1675056114 start_time against this md array: /dev/md1 state is clean, 1675056119 start_time against this md array: /dev/md1 state is clean, 1675056124 start_time against this md array: /dev/md1 state is clean, 1675056130 start_time against this md array: /dev/md1 state is clean, 1675056135 start_time against this md array: /dev/md1 state is clean, 1675056140 start_time against this md array: /dev/md1 state is clean, 1675056145 start_time against this md array: /dev/md1 state is clean, 1675056151 start_time against this md array: /dev/md1 state is clean, 1675056156 start_time against this md array: /dev/md1 state is clean :: [ 00:22:36 ] :: [ LOG ] :: mkfs -t xfs /dev/md1 log stripe unit (524288 bytes) is too large (maximum is 256KiB) log stripe unit adjusted to 32KiB meta-data=/dev/md1 isize=512 agcount=8, agsize=63744 blks = sectsz=512 attr=2, projid32bit=1 = crc=1 finobt=1, sparse=1, rmapbt=0 = reflink=1 bigtime=1 inobtcount=1 data = bsize=4096 blocks=509952, imaxpct=25 = sunit=128 swidth=512 blks naming =version 2 bsize=4096 ascii-ci=0, ftype=1 log =internal log bsize=4096 blocks=2560, version=2 = sectsz=512 sunit=8 blks, lazy-count=1 realtime =none extsz=4096 blocks=0, rtextents=0 Discarding blocks...Done. :: [ 00:23:55 ] :: [ BEGIN ] :: Running 'mount -t xfs /dev/md1 /mnt/md_test ' :: [ 00:23:55 ] :: [ PASS ] :: Command 'mount -t xfs /dev/md1 /mnt/md_test ' (Expected 0, got 0) :: [ 00:23:56 ] :: [ BEGIN ] :: Running 'fstrim -v /mnt/md_test' /mnt/md_test: 1.9 GiB (2077425664 bytes) trimmed :: [ 00:24:36 ] :: [ PASS ] :: Command 'fstrim -v /mnt/md_test' (Expected 0, got 0) :: [ 00:24:36 ] :: [ BEGIN ] :: Running 'umount /dev/md1' :: [ 00:24:37 ] :: [ PASS ] :: Command 'umount /dev/md1' (Expected 0, got 0) INFO: Executing MD_Clean_RAID() against this md device: mdadm --stop /dev/md1 mdadm: stopped /dev/md1 clean devs : /dev/loop0 /dev/loop3 /dev/loop1 /dev/loop2 /dev/loop4 /dev/loop5 mdadm --zero-superblock /dev/loop0 mdadm --zero-superblock /dev/loop3 mdadm --zero-superblock /dev/loop1 mdadm --zero-superblock /dev/loop2 mdadm --zero-superblock /dev/loop4 mdadm --zero-superblock /dev/loop5 ret is 0 ls /dev/md1 ls: cannot access '/dev/md1': No such file or directory mdadm --stop can delete md node name /dev/md1 in /dev INFO: Executing MD_Create_RAID() to create raid 5 INFO: Created md raid with these raid devices " /dev/loop0 /dev/loop3 /dev/loop1 /dev/loop2 /dev/loop4". INFO: Created md raid with these spare disks " /dev/loop5". :: [ 00:25:04 ] :: [ BEGIN ] :: Running 'mdadm --create --run /dev/md1 --level 5 --metadata 1.2 --raid-devices 5 /dev/loop0 /dev/loop3 /dev/loop1 /dev/loop2 /dev/loop4 --spare-devices 1 /dev/loop5 --chunk 512 --bitmap=internal --bitmap-chunk=64M' mdadm: array /dev/md1 started. :: [ 00:25:05 ] :: [ PASS ] :: Command 'mdadm --create --run /dev/md1 --level 5 --metadata 1.2 --raid-devices 5 /dev/loop0 /dev/loop3 /dev/loop1 /dev/loop2 /dev/loop4 --spare-devices 1 /dev/loop5 --chunk 512 --bitmap=internal --bitmap-chunk=64M' (Expected 0, got 0) INFO:cat /proc/mdstat###################### :: [ 00:25:05 ] :: [ BEGIN ] :: Running 'cat /proc/mdstat' Personalities : [raid6] [raid5] [raid4] [raid0] [raid1] md1 : active raid5 loop4[6] loop5[5](S) loop2[3] loop1[2] loop3[1] loop0[0] 2039808 blocks super 1.2 level 5, 512k chunk, algorithm 2 [5/4] [UUUU_] [>....................] recovery = 0.9% (4908/509952) finish=1.7min speed=4908K/sec bitmap: 1/1 pages [4KB], 65536KB chunk unused devices: :: [ 00:25:05 ] :: [ PASS ] :: Command 'cat /proc/mdstat' (Expected 0, got 0) :: [ 00:25:06 ] :: [ BEGIN ] :: Running 'lsblk' NAME MAJ:MIN RM SIZE RO TYPE MOUNTPOINTS loop0 7:0 0 500M 0 loop └─md1 9:1 0 1.9G 0 raid5 loop1 7:1 0 500M 0 loop └─md1 9:1 0 1.9G 0 raid5 loop2 7:2 0 500M 0 loop └─md1 9:1 0 1.9G 0 raid5 loop3 7:3 0 500M 0 loop └─md1 9:1 0 1.9G 0 raid5 loop4 7:4 0 500M 0 loop └─md1 9:1 0 1.9G 0 raid5 loop5 7:5 0 500M 0 loop └─md1 9:1 0 1.9G 0 raid5 sda 8:0 0 465.7G 0 disk ├─sda1 8:1 0 1G 0 part /boot └─sda2 8:2 0 464.7G 0 part ├─cs_hpe--dl360pgen8--08-root 253:0 0 70G 0 lvm / ├─cs_hpe--dl360pgen8--08-swap 253:1 0 15.7G 0 lvm [SWAP] └─cs_hpe--dl360pgen8--08-home 253:2 0 379G 0 lvm /home :: [ 00:25:06 ] :: [ PASS ] :: Command 'lsblk' (Expected 0, got 0) /dev/md1 INFO:mdadm -D /dev/md1 ######################### :: [ 00:25:06 ] :: [ BEGIN ] :: Running 'mdadm --detail /dev/md1' /dev/md1: Version : 1.2 Creation Time : Mon Jan 30 00:25:04 2023 Raid Level : raid5 Array Size : 2039808 (1992.00 MiB 2088.76 MB) Used Dev Size : 509952 (498.00 MiB 522.19 MB) Raid Devices : 5 Total Devices : 6 Persistence : Superblock is persistent Intent Bitmap : Internal Update Time : Mon Jan 30 00:25:05 2023 State : clean, degraded, recovering Active Devices : 4 Working Devices : 6 Failed Devices : 0 Spare Devices : 2 Layout : left-symmetric Chunk Size : 512K Consistency Policy : bitmap Rebuild Status : 4% complete Name : 1 UUID : b13d1c14:7fb01a2e:02ed4550:4f1c397f Events : 1 Number Major Minor RaidDevice State 0 7 0 0 active sync /dev/loop0 1 7 3 1 active sync /dev/loop3 2 7 1 2 active sync /dev/loop1 3 7 2 3 active sync /dev/loop2 6 7 4 4 spare rebuilding /dev/loop4 5 7 5 - spare /dev/loop5 :: [ 00:25:07 ] :: [ PASS ] :: Command 'mdadm --detail /dev/md1' (Expected 0, got 0) :: [ 00:25:07 ] :: [ LOG ] :: INFO: Successfully created md raid /dev/md1 1675056307 start_time against this md array: /dev/md1 state is clean, 1675056312 start_time against this md array: /dev/md1 state is clean, 1675056317 start_time against this md array: /dev/md1 state is clean, 1675056323 start_time against this md array: /dev/md1 state is clean, 1675056328 start_time against this md array: /dev/md1 state is clean, 1675056333 start_time against this md array: /dev/md1 state is clean, 1675056338 start_time against this md array: /dev/md1 state is clean, 1675056344 start_time against this md array: /dev/md1 state is clean, 1675056349 start_time against this md array: /dev/md1 state is clean, 1675056354 start_time against this md array: /dev/md1 state is clean :: [ 00:25:54 ] :: [ LOG ] :: mkfs -t xfs /dev/md1 mkfs.xfs: /dev/md1 appears to contain an existing filesystem (xfs). mkfs.xfs: Use the -f option to force overwrite. log stripe unit (524288 bytes) is too large (maximum is 256KiB) log stripe unit adjusted to 32KiB meta-data=/dev/md1 isize=512 agcount=8, agsize=63744 blks = sectsz=512 attr=2, projid32bit=1 = crc=1 finobt=1, sparse=1, rmapbt=0 = reflink=1 bigtime=1 inobtcount=1 data = bsize=4096 blocks=509952, imaxpct=25 = sunit=128 swidth=512 blks naming =version 2 bsize=4096 ascii-ci=0, ftype=1 log =internal log bsize=4096 blocks=2560, version=2 = sectsz=512 sunit=8 blks, lazy-count=1 realtime =none extsz=4096 blocks=0, rtextents=0 Discarding blocks...Done. :: [ 00:27:12 ] :: [ BEGIN ] :: Running 'mount -t xfs /dev/md1 /mnt/md_test ' :: [ 00:27:12 ] :: [ PASS ] :: Command 'mount -t xfs /dev/md1 /mnt/md_test ' (Expected 0, got 0) :: [ 00:27:13 ] :: [ BEGIN ] :: Running 'fstrim -v /mnt/md_test' /mnt/md_test: 1.9 GiB (2077425664 bytes) trimmed :: [ 00:27:52 ] :: [ PASS ] :: Command 'fstrim -v /mnt/md_test' (Expected 0, got 0) :: [ 00:27:53 ] :: [ BEGIN ] :: Running 'umount /dev/md1' :: [ 00:27:53 ] :: [ PASS ] :: Command 'umount /dev/md1' (Expected 0, got 0) INFO: Executing MD_Clean_RAID() against this md device: mdadm --stop /dev/md1 mdadm: stopped /dev/md1 clean devs : /dev/loop0 /dev/loop3 /dev/loop1 /dev/loop2 /dev/loop4 /dev/loop5 mdadm --zero-superblock /dev/loop0 mdadm --zero-superblock /dev/loop3 mdadm --zero-superblock /dev/loop1 mdadm --zero-superblock /dev/loop2 mdadm --zero-superblock /dev/loop4 mdadm --zero-superblock /dev/loop5 ret is 0 ls /dev/md1 ls: cannot access '/dev/md1': No such file or directory mdadm --stop can delete md node name /dev/md1 in /dev INFO: Executing MD_Create_RAID() to create raid 6 INFO: Created md raid with these raid devices " /dev/loop0 /dev/loop3 /dev/loop1 /dev/loop2 /dev/loop4". INFO: Created md raid with these spare disks " /dev/loop5". :: [ 00:28:20 ] :: [ BEGIN ] :: Running 'mdadm --create --run /dev/md1 --level 6 --metadata 1.2 --raid-devices 5 /dev/loop0 /dev/loop3 /dev/loop1 /dev/loop2 /dev/loop4 --spare-devices 1 /dev/loop5 --chunk 512 --bitmap=internal --bitmap-chunk=64M' mdadm: array /dev/md1 started. :: [ 00:28:21 ] :: [ PASS ] :: Command 'mdadm --create --run /dev/md1 --level 6 --metadata 1.2 --raid-devices 5 /dev/loop0 /dev/loop3 /dev/loop1 /dev/loop2 /dev/loop4 --spare-devices 1 /dev/loop5 --chunk 512 --bitmap=internal --bitmap-chunk=64M' (Expected 0, got 0) INFO:cat /proc/mdstat###################### :: [ 00:28:21 ] :: [ BEGIN ] :: Running 'cat /proc/mdstat' Personalities : [raid6] [raid5] [raid4] [raid0] [raid1] md1 : active raid6 loop5[5](S) loop4[4] loop2[3] loop1[2] loop3[1] loop0[0] 1529856 blocks super 1.2 level 6, 512k chunk, algorithm 2 [5/5] [UUUUU] [>....................] resync = 1.5% (8012/509952) finish=1.0min speed=8012K/sec bitmap: 1/1 pages [4KB], 65536KB chunk unused devices: :: [ 00:28:21 ] :: [ PASS ] :: Command 'cat /proc/mdstat' (Expected 0, got 0) :: [ 00:28:22 ] :: [ BEGIN ] :: Running 'lsblk' NAME MAJ:MIN RM SIZE RO TYPE MOUNTPOINTS loop0 7:0 0 500M 0 loop └─md1 9:1 0 1.5G 0 raid6 loop1 7:1 0 500M 0 loop └─md1 9:1 0 1.5G 0 raid6 loop2 7:2 0 500M 0 loop └─md1 9:1 0 1.5G 0 raid6 loop3 7:3 0 500M 0 loop └─md1 9:1 0 1.5G 0 raid6 loop4 7:4 0 500M 0 loop └─md1 9:1 0 1.5G 0 raid6 loop5 7:5 0 500M 0 loop └─md1 9:1 0 1.5G 0 raid6 sda 8:0 0 465.7G 0 disk ├─sda1 8:1 0 1G 0 part /boot └─sda2 8:2 0 464.7G 0 part ├─cs_hpe--dl360pgen8--08-root 253:0 0 70G 0 lvm / ├─cs_hpe--dl360pgen8--08-swap 253:1 0 15.7G 0 lvm [SWAP] └─cs_hpe--dl360pgen8--08-home 253:2 0 379G 0 lvm /home :: [ 00:28:22 ] :: [ PASS ] :: Command 'lsblk' (Expected 0, got 0) /dev/md1 INFO:mdadm -D /dev/md1 ######################### :: [ 00:28:22 ] :: [ BEGIN ] :: Running 'mdadm --detail /dev/md1' /dev/md1: Version : 1.2 Creation Time : Mon Jan 30 00:28:20 2023 Raid Level : raid6 Array Size : 1529856 (1494.00 MiB 1566.57 MB) Used Dev Size : 509952 (498.00 MiB 522.19 MB) Raid Devices : 5 Total Devices : 6 Persistence : Superblock is persistent Intent Bitmap : Internal Update Time : Mon Jan 30 00:28:20 2023 State : clean, resyncing Active Devices : 5 Working Devices : 6 Failed Devices : 0 Spare Devices : 1 Layout : left-symmetric Chunk Size : 512K Consistency Policy : bitmap Resync Status : 5% complete Name : 1 UUID : 4b2a7145:ccabe869:95e69726:f0862ac5 Events : 0 Number Major Minor RaidDevice State 0 7 0 0 active sync /dev/loop0 1 7 3 1 active sync /dev/loop3 2 7 1 2 active sync /dev/loop1 3 7 2 3 active sync /dev/loop2 4 7 4 4 active sync /dev/loop4 5 7 5 - spare /dev/loop5 :: [ 00:28:23 ] :: [ PASS ] :: Command 'mdadm --detail /dev/md1' (Expected 0, got 0) :: [ 00:28:23 ] :: [ LOG ] :: INFO: Successfully created md raid /dev/md1 1675056503 start_time against this md array: /dev/md1 state is clean, 1675056508 start_time against this md array: /dev/md1 state is clean, 1675056514 start_time against this md array: /dev/md1 state is clean, 1675056519 start_time against this md array: /dev/md1 state is clean, 1675056524 start_time against this md array: /dev/md1 state is clean, 1675056529 start_time against this md array: /dev/md1 state is clean, 1675056534 start_time against this md array: /dev/md1 state is clean :: [ 00:28:55 ] :: [ LOG ] :: mkfs -t xfs /dev/md1 log stripe unit (524288 bytes) is too large (maximum is 256KiB) log stripe unit adjusted to 32KiB meta-data=/dev/md1 isize=512 agcount=8, agsize=47872 blks = sectsz=512 attr=2, projid32bit=1 = crc=1 finobt=1, sparse=1, rmapbt=0 = reflink=1 bigtime=1 inobtcount=1 data = bsize=4096 blocks=382464, imaxpct=25 = sunit=128 swidth=384 blks naming =version 2 bsize=4096 ascii-ci=0, ftype=1 log =internal log bsize=4096 blocks=2560, version=2 = sectsz=512 sunit=8 blks, lazy-count=1 realtime =none extsz=4096 blocks=0, rtextents=0 Discarding blocks...Done. :: [ 00:29:41 ] :: [ BEGIN ] :: Running 'mount -t xfs /dev/md1 /mnt/md_test ' :: [ 00:29:41 ] :: [ PASS ] :: Command 'mount -t xfs /dev/md1 /mnt/md_test ' (Expected 0, got 0) :: [ 00:29:42 ] :: [ BEGIN ] :: Running 'fstrim -v /mnt/md_test' /mnt/md_test: 1.4 GiB (1555234816 bytes) trimmed :: [ 00:30:17 ] :: [ PASS ] :: Command 'fstrim -v /mnt/md_test' (Expected 0, got 0) :: [ 00:30:17 ] :: [ BEGIN ] :: Running 'umount /dev/md1' :: [ 00:30:18 ] :: [ PASS ] :: Command 'umount /dev/md1' (Expected 0, got 0) INFO: Executing MD_Clean_RAID() against this md device: mdadm --stop /dev/md1 mdadm: stopped /dev/md1 clean devs : /dev/loop0 /dev/loop3 /dev/loop1 /dev/loop2 /dev/loop4 /dev/loop5 mdadm --zero-superblock /dev/loop0 mdadm --zero-superblock /dev/loop3 mdadm --zero-superblock /dev/loop1 mdadm --zero-superblock /dev/loop2 mdadm --zero-superblock /dev/loop4 mdadm --zero-superblock /dev/loop5 ret is 0 ls /dev/md1 ls: cannot access '/dev/md1': No such file or directory mdadm --stop can delete md node name /dev/md1 in /dev INFO: Executing MD_Create_RAID() to create raid 10 INFO: Created md raid with these raid devices " /dev/loop0 /dev/loop3 /dev/loop1 /dev/loop2 /dev/loop4". INFO: Created md raid with these spare disks " /dev/loop5". :: [ 00:30:45 ] :: [ BEGIN ] :: Running 'mdadm --create --run /dev/md1 --level 10 --metadata 1.2 --raid-devices 5 /dev/loop0 /dev/loop3 /dev/loop1 /dev/loop2 /dev/loop4 --spare-devices 1 /dev/loop5 --chunk 512 --bitmap=internal --bitmap-chunk=64M' mdadm: array /dev/md1 started. :: [ 00:30:46 ] :: [ PASS ] :: Command 'mdadm --create --run /dev/md1 --level 10 --metadata 1.2 --raid-devices 5 /dev/loop0 /dev/loop3 /dev/loop1 /dev/loop2 /dev/loop4 --spare-devices 1 /dev/loop5 --chunk 512 --bitmap=internal --bitmap-chunk=64M' (Expected 0, got 0) INFO:cat /proc/mdstat###################### :: [ 00:30:46 ] :: [ BEGIN ] :: Running 'cat /proc/mdstat' Personalities : [raid6] [raid5] [raid4] [raid0] [raid1] [raid10] md1 : active raid10 loop5[5](S) loop4[4] loop2[3] loop1[2] loop3[1] loop0[0] 1274880 blocks super 1.2 512K chunks 2 near-copies [5/5] [UUUUU] [>....................] resync = 2.4% (32128/1274880) finish=0.6min speed=32128K/sec bitmap: 1/1 pages [4KB], 65536KB chunk unused devices: :: [ 00:30:46 ] :: [ PASS ] :: Command 'cat /proc/mdstat' (Expected 0, got 0) :: [ 00:30:46 ] :: [ BEGIN ] :: Running 'lsblk' NAME MAJ:MIN RM SIZE RO TYPE MOUNTPOINTS loop0 7:0 0 500M 0 loop └─md1 9:1 0 1.2G 0 raid10 loop1 7:1 0 500M 0 loop └─md1 9:1 0 1.2G 0 raid10 loop2 7:2 0 500M 0 loop └─md1 9:1 0 1.2G 0 raid10 loop3 7:3 0 500M 0 loop └─md1 9:1 0 1.2G 0 raid10 loop4 7:4 0 500M 0 loop └─md1 9:1 0 1.2G 0 raid10 loop5 7:5 0 500M 0 loop └─md1 9:1 0 1.2G 0 raid10 sda 8:0 0 465.7G 0 disk ├─sda1 8:1 0 1G 0 part /boot └─sda2 8:2 0 464.7G 0 part ├─cs_hpe--dl360pgen8--08-root 253:0 0 70G 0 lvm / ├─cs_hpe--dl360pgen8--08-swap 253:1 0 15.7G 0 lvm [SWAP] └─cs_hpe--dl360pgen8--08-home 253:2 0 379G 0 lvm /home :: [ 00:30:47 ] :: [ PASS ] :: Command 'lsblk' (Expected 0, got 0) /dev/md1 INFO:mdadm -D /dev/md1 ######################### :: [ 00:30:47 ] :: [ BEGIN ] :: Running 'mdadm --detail /dev/md1' /dev/md1: Version : 1.2 Creation Time : Mon Jan 30 00:30:45 2023 Raid Level : raid10 Array Size : 1274880 (1245.00 MiB 1305.48 MB) Used Dev Size : 509952 (498.00 MiB 522.19 MB) Raid Devices : 5 Total Devices : 6 Persistence : Superblock is persistent Intent Bitmap : Internal Update Time : Mon Jan 30 00:30:47 2023 State : clean, resyncing Active Devices : 5 Working Devices : 6 Failed Devices : 0 Spare Devices : 1 Layout : near=2 Chunk Size : 512K Consistency Policy : bitmap Resync Status : 12% complete Name : 1 UUID : e2d70fed:5e7c0ae8:4e642b07:758f7326 Events : 1 Number Major Minor RaidDevice State 0 7 0 0 active sync /dev/loop0 1 7 3 1 active sync /dev/loop3 2 7 1 2 active sync /dev/loop1 3 7 2 3 active sync /dev/loop2 4 7 4 4 active sync /dev/loop4 5 7 5 - spare /dev/loop5 :: [ 00:30:48 ] :: [ PASS ] :: Command 'mdadm --detail /dev/md1' (Expected 0, got 0) :: [ 00:30:48 ] :: [ LOG ] :: INFO: Successfully created md raid /dev/md1 1675056648 start_time against this md array: /dev/md1 state is clean, 1675056653 start_time against this md array: /dev/md1 state is clean, 1675056659 start_time against this md array: /dev/md1 state is clean :: [ 00:30:59 ] :: [ LOG ] :: mkfs -t xfs /dev/md1 log stripe unit (524288 bytes) is too large (maximum is 256KiB) log stripe unit adjusted to 32KiB meta-data=/dev/md1 isize=512 agcount=8, agsize=39936 blks = sectsz=512 attr=2, projid32bit=1 = crc=1 finobt=1, sparse=1, rmapbt=0 = reflink=1 bigtime=1 inobtcount=1 data = bsize=4096 blocks=318720, imaxpct=25 = sunit=128 swidth=640 blks naming =version 2 bsize=4096 ascii-ci=0, ftype=1 log =internal log bsize=4096 blocks=2560, version=2 = sectsz=512 sunit=8 blks, lazy-count=1 realtime =none extsz=4096 blocks=0, rtextents=0 Discarding blocks...Done. :: [ 00:31:01 ] :: [ BEGIN ] :: Running 'mount -t xfs /dev/md1 /mnt/md_test ' :: [ 00:31:02 ] :: [ PASS ] :: Command 'mount -t xfs /dev/md1 /mnt/md_test ' (Expected 0, got 0) :: [ 00:31:02 ] :: [ BEGIN ] :: Running 'fstrim -v /mnt/md_test' /mnt/md_test: 1.2 GiB (1294630912 bytes) trimmed :: [ 00:31:02 ] :: [ PASS ] :: Command 'fstrim -v /mnt/md_test' (Expected 0, got 0) :: [ 00:31:03 ] :: [ BEGIN ] :: Running 'umount /dev/md1' :: [ 00:31:03 ] :: [ PASS ] :: Command 'umount /dev/md1' (Expected 0, got 0) INFO: Executing MD_Clean_RAID() against this md device: mdadm --stop /dev/md1 mdadm: stopped /dev/md1 clean devs : /dev/loop0 /dev/loop3 /dev/loop1 /dev/loop2 /dev/loop4 /dev/loop5 mdadm --zero-superblock /dev/loop0 mdadm --zero-superblock /dev/loop3 mdadm --zero-superblock /dev/loop1 mdadm --zero-superblock /dev/loop2 mdadm --zero-superblock /dev/loop4 mdadm --zero-superblock /dev/loop5 ret is 0 ls /dev/md1 ls: cannot access '/dev/md1': No such file or directory mdadm --stop can delete md node name /dev/md1 in /dev losetup -d /dev/loop0 losetup -d /dev/loop3 losetup -d /dev/loop1 losetup -d /dev/loop2 losetup -d /dev/loop4 losetup -d /dev/loop5 :: [ 00:31:25 ] :: [ BEGIN ] :: check the errors :: actually running 'dmesg | grep -i 'Call Trace:'' :: [ 00:31:26 ] :: [ PASS ] :: check the errors (Expected 1, got 1) :::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::: :: Duration: 771s :: Assertions: 47 good, 0 bad :: RESULT: PASS (Test) ** Test PASS Score:0 Uploading resultoutputfile.log .done :::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::: :: TEST PROTOCOL :::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::: Test run ID : 9 Package : unknown beakerlib RPM : beakerlib-1.29.2-2.el9.noarch bl-redhat RPM : beakerlib-redhat-1-33.el9.noarch Test name : unknown Test started : 2023-01-30 00:18:33 EST Test finished : 2023-01-30 00:31:31 EST (still running) Test duration : 778 seconds Distro : CentOS Stream release 9 Hostname : hpe-dl360pgen8-08.hpe2.lab.eng.bos.redhat.com Architecture : x86_64 CPUs : 24 x Intel(R) Xeon(R) CPU E5-2620 v2 @ 2.10GHz RAM size : 29885 MB HDD size : 449.75 GB :::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::: :: Test description :::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::: PURPOSE of trim Description: Test the function of TRIM which quickly erase invalid data for the RAID that created by mdadm. :::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::: :: Test :::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::: :: [ 00:18:36 ] :: [ PASS ] :: Command 'uname -a' (Expected 0, got 0) :: [ 00:18:38 ] :: [ PASS ] :: Command 'rpm -q mdadm || dnf install -y mdadm' (Expected 0, got 0) :: [ 00:18:38 ] :: [ LOG ] :: ./runtest.sh :: [ 00:18:39 ] :: [ PASS ] :: Command 'modprobe raid456 devices_handle_discard_safely=Y' (Expected 0, got 0) :: [ 00:18:39 ] :: [ PASS ] :: Command 'echo Y >/sys/module/raid456/parameters/devices_handle_discard_safely' (Expected 0, got 0) :: [ 00:19:35 ] :: [ PASS ] :: Command 'mdadm --create --run /dev/md1 --level 0 --metadata 1.2 --raid-devices 5 /dev/loop0 /dev/loop3 /dev/loop1 /dev/loop2 /dev/loop4 --chunk 512' (Expected 0, got 0) :: [ 00:19:36 ] :: [ PASS ] :: Command 'cat /proc/mdstat' (Expected 0, got 0) :: [ 00:19:36 ] :: [ PASS ] :: Command 'lsblk' (Expected 0, got 0) :: [ 00:19:37 ] :: [ PASS ] :: Command 'mdadm --detail /dev/md1' (Expected 0, got 0) :: [ 00:19:37 ] :: [ LOG ] :: INFO: Successfully created md raid /dev/md1 :: [ 00:19:38 ] :: [ LOG ] :: mkfs -t xfs /dev/md1 :: [ 00:19:40 ] :: [ PASS ] :: Command 'mount -t xfs /dev/md1 /mnt/md_test ' (Expected 0, got 0) :: [ 00:19:41 ] :: [ PASS ] :: Command 'fstrim -v /mnt/md_test' (Expected 0, got 0) :: [ 00:19:42 ] :: [ PASS ] :: Command 'umount /dev/md1' (Expected 0, got 0) :: [ 00:20:09 ] :: [ PASS ] :: Command 'mdadm --create --run /dev/md1 --level 1 --metadata 1.2 --raid-devices 5 /dev/loop0 /dev/loop3 /dev/loop1 /dev/loop2 /dev/loop4 --spare-devices 1 /dev/loop5 --bitmap=internal --bitmap-chunk=64M' (Expected 0, got 0) :: [ 00:20:10 ] :: [ PASS ] :: Command 'cat /proc/mdstat' (Expected 0, got 0) :: [ 00:20:10 ] :: [ PASS ] :: Command 'lsblk' (Expected 0, got 0) :: [ 00:20:11 ] :: [ PASS ] :: Command 'mdadm --detail /dev/md1' (Expected 0, got 0) :: [ 00:20:11 ] :: [ LOG ] :: INFO: Successfully created md raid /dev/md1 :: [ 00:21:14 ] :: [ LOG ] :: mkfs -t xfs /dev/md1 :: [ 00:21:17 ] :: [ PASS ] :: Command 'mount -t xfs /dev/md1 /mnt/md_test ' (Expected 0, got 0) :: [ 00:21:18 ] :: [ PASS ] :: Command 'fstrim -v /mnt/md_test' (Expected 0, got 0) :: [ 00:21:19 ] :: [ PASS ] :: Command 'umount /dev/md1' (Expected 0, got 0) :: [ 00:21:47 ] :: [ PASS ] :: Command 'mdadm --create --run /dev/md1 --level 4 --metadata 1.2 --raid-devices 5 /dev/loop0 /dev/loop3 /dev/loop1 /dev/loop2 /dev/loop4 --spare-devices 1 /dev/loop5 --chunk 512 --bitmap=internal --bitmap-chunk=64M' (Expected 0, got 0) :: [ 00:21:47 ] :: [ PASS ] :: Command 'cat /proc/mdstat' (Expected 0, got 0) :: [ 00:21:48 ] :: [ PASS ] :: Command 'lsblk' (Expected 0, got 0) :: [ 00:21:48 ] :: [ PASS ] :: Command 'mdadm --detail /dev/md1' (Expected 0, got 0) :: [ 00:21:49 ] :: [ LOG ] :: INFO: Successfully created md raid /dev/md1 :: [ 00:22:36 ] :: [ LOG ] :: mkfs -t xfs /dev/md1 :: [ 00:23:55 ] :: [ PASS ] :: Command 'mount -t xfs /dev/md1 /mnt/md_test ' (Expected 0, got 0) :: [ 00:24:36 ] :: [ PASS ] :: Command 'fstrim -v /mnt/md_test' (Expected 0, got 0) :: [ 00:24:37 ] :: [ PASS ] :: Command 'umount /dev/md1' (Expected 0, got 0) :: [ 00:25:05 ] :: [ PASS ] :: Command 'mdadm --create --run /dev/md1 --level 5 --metadata 1.2 --raid-devices 5 /dev/loop0 /dev/loop3 /dev/loop1 /dev/loop2 /dev/loop4 --spare-devices 1 /dev/loop5 --chunk 512 --bitmap=internal --bitmap-chunk=64M' (Expected 0, got 0) :: [ 00:25:05 ] :: [ PASS ] :: Command 'cat /proc/mdstat' (Expected 0, got 0) :: [ 00:25:06 ] :: [ PASS ] :: Command 'lsblk' (Expected 0, got 0) :: [ 00:25:07 ] :: [ PASS ] :: Command 'mdadm --detail /dev/md1' (Expected 0, got 0) :: [ 00:25:07 ] :: [ LOG ] :: INFO: Successfully created md raid /dev/md1 :: [ 00:25:54 ] :: [ LOG ] :: mkfs -t xfs /dev/md1 :: [ 00:27:12 ] :: [ PASS ] :: Command 'mount -t xfs /dev/md1 /mnt/md_test ' (Expected 0, got 0) :: [ 00:27:52 ] :: [ PASS ] :: Command 'fstrim -v /mnt/md_test' (Expected 0, got 0) :: [ 00:27:53 ] :: [ PASS ] :: Command 'umount /dev/md1' (Expected 0, got 0) :: [ 00:28:21 ] :: [ PASS ] :: Command 'mdadm --create --run /dev/md1 --level 6 --metadata 1.2 --raid-devices 5 /dev/loop0 /dev/loop3 /dev/loop1 /dev/loop2 /dev/loop4 --spare-devices 1 /dev/loop5 --chunk 512 --bitmap=internal --bitmap-chunk=64M' (Expected 0, got 0) :: [ 00:28:21 ] :: [ PASS ] :: Command 'cat /proc/mdstat' (Expected 0, got 0) :: [ 00:28:22 ] :: [ PASS ] :: Command 'lsblk' (Expected 0, got 0) :: [ 00:28:23 ] :: [ PASS ] :: Command 'mdadm --detail /dev/md1' (Expected 0, got 0) :: [ 00:28:23 ] :: [ LOG ] :: INFO: Successfully created md raid /dev/md1 :: [ 00:28:55 ] :: [ LOG ] :: mkfs -t xfs /dev/md1 :: [ 00:29:41 ] :: [ PASS ] :: Command 'mount -t xfs /dev/md1 /mnt/md_test ' (Expected 0, got 0) :: [ 00:30:17 ] :: [ PASS ] :: Command 'fstrim -v /mnt/md_test' (Expected 0, got 0) :: [ 00:30:18 ] :: [ PASS ] :: Command 'umount /dev/md1' (Expected 0, got 0) :: [ 00:30:46 ] :: [ PASS ] :: Command 'mdadm --create --run /dev/md1 --level 10 --metadata 1.2 --raid-devices 5 /dev/loop0 /dev/loop3 /dev/loop1 /dev/loop2 /dev/loop4 --spare-devices 1 /dev/loop5 --chunk 512 --bitmap=internal --bitmap-chunk=64M' (Expected 0, got 0) :: [ 00:30:46 ] :: [ PASS ] :: Command 'cat /proc/mdstat' (Expected 0, got 0) :: [ 00:30:47 ] :: [ PASS ] :: Command 'lsblk' (Expected 0, got 0) :: [ 00:30:48 ] :: [ PASS ] :: Command 'mdadm --detail /dev/md1' (Expected 0, got 0) :: [ 00:30:48 ] :: [ LOG ] :: INFO: Successfully created md raid /dev/md1 :: [ 00:30:59 ] :: [ LOG ] :: mkfs -t xfs /dev/md1 :: [ 00:31:02 ] :: [ PASS ] :: Command 'mount -t xfs /dev/md1 /mnt/md_test ' (Expected 0, got 0) :: [ 00:31:02 ] :: [ PASS ] :: Command 'fstrim -v /mnt/md_test' (Expected 0, got 0) :: [ 00:31:03 ] :: [ PASS ] :: Command 'umount /dev/md1' (Expected 0, got 0) :: [ 00:31:26 ] :: [ PASS ] :: check the errors (Expected 1, got 1) :::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::: :: Duration: 771s :: Assertions: 47 good, 0 bad :: RESULT: PASS (Test) :::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::: :: unknown :::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::: :: [ 00:31:32 ] :: [ LOG ] :: Phases fingerprint: L5rLAvqh :: [ 00:31:33 ] :: [ LOG ] :: Asserts fingerprint: IcvkrVvU Uploading journal.xml .done :::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::: :: Duration: 779s :: Phases: 1 good, 0 bad :: OVERALL RESULT: PASS (unknown)