use_pty:FALSE /usr/share/restraint/plugins/run_task_plugins bash ./runtest.sh :::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::: :: Test :::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::: :: [ 15:54:48 ] :: [ BEGIN ] :: Running 'uname -a' Linux hpe-ml350gen9-01.hpe2.lab.eng.bos.redhat.com 5.14.0-246.rt14.245.1955_759844798.el9.x86_64+debug #1 SMP PREEMPT_RT Fri Jan 27 16:35:10 UTC 2023 x86_64 x86_64 x86_64 GNU/Linux :: [ 15:54:48 ] :: [ PASS ] :: Command 'uname -a' (Expected 0, got 0) :: [ 15:54:49 ] :: [ BEGIN ] :: Running 'rpm -q mdadm || dnf install -y mdadm' mdadm-4.2-8.el9.x86_64 :: [ 15:54:50 ] :: [ PASS ] :: Command 'rpm -q mdadm || dnf install -y mdadm' (Expected 0, got 0) :: [ 15:54:50 ] :: [ LOG ] :: ./runtest.sh :: [ 15:54:50 ] :: [ BEGIN ] :: Running 'modprobe raid456 devices_handle_discard_safely=Y' :: [ 15:54:50 ] :: [ PASS ] :: Command 'modprobe raid456 devices_handle_discard_safely=Y' (Expected 0, got 0) :: [ 15:54:51 ] :: [ BEGIN ] :: Running 'echo Y >/sys/module/raid456/parameters/devices_handle_discard_safely' :: [ 15:54:51 ] :: [ PASS ] :: Command 'echo Y >/sys/module/raid456/parameters/devices_handle_discard_safely' (Expected 0, got 0) /usr/sbin/mkfs.xfs INFO: Executing MD_Create_RAID() to create raid 0 INFO: Created md raid with these raid devices " /dev/loop0 /dev/loop3 /dev/loop1 /dev/loop2 /dev/loop4". :: [ 15:55:11 ] :: [ BEGIN ] :: Running 'mdadm --create --run /dev/md1 --level 0 --metadata 1.2 --raid-devices 5 /dev/loop0 /dev/loop3 /dev/loop1 /dev/loop2 /dev/loop4 --chunk 512' mdadm: array /dev/md1 started. :: [ 15:55:18 ] :: [ PASS ] :: Command 'mdadm --create --run /dev/md1 --level 0 --metadata 1.2 --raid-devices 5 /dev/loop0 /dev/loop3 /dev/loop1 /dev/loop2 /dev/loop4 --chunk 512' (Expected 0, got 0) INFO:cat /proc/mdstat###################### :: [ 15:55:18 ] :: [ BEGIN ] :: Running 'cat /proc/mdstat' Personalities : [raid6] [raid5] [raid4] [raid0] md1 : active raid0 loop4[4] loop2[3] loop1[2] loop3[1] loop0[0] 2549760 blocks super 1.2 512k chunks unused devices: :: [ 15:55:19 ] :: [ PASS ] :: Command 'cat /proc/mdstat' (Expected 0, got 0) :: [ 15:55:19 ] :: [ BEGIN ] :: Running 'lsblk' NAME MAJ:MIN RM SIZE RO TYPE MOUNTPOINTS loop0 7:0 0 500M 0 loop └─md1 9:1 0 2.4G 0 raid0 loop1 7:1 0 500M 0 loop └─md1 9:1 0 2.4G 0 raid0 loop2 7:2 0 500M 0 loop └─md1 9:1 0 2.4G 0 raid0 loop3 7:3 0 500M 0 loop └─md1 9:1 0 2.4G 0 raid0 loop4 7:4 0 500M 0 loop └─md1 9:1 0 2.4G 0 raid0 loop5 7:5 0 500M 0 loop sda 8:0 0 558.7G 0 disk ├─sda1 8:1 0 1G 0 part /boot └─sda2 8:2 0 557.7G 0 part ├─cs_hpe--ml350gen9--01-root 253:0 0 70G 0 lvm / ├─cs_hpe--ml350gen9--01-swap 253:1 0 23.6G 0 lvm [SWAP] └─cs_hpe--ml350gen9--01-home 253:2 0 464.2G 0 lvm /home :: [ 15:55:19 ] :: [ PASS ] :: Command 'lsblk' (Expected 0, got 0) /dev/md1 INFO:mdadm -D /dev/md1 ######################### :: [ 15:55:19 ] :: [ BEGIN ] :: Running 'mdadm --detail /dev/md1' /dev/md1: Version : 1.2 Creation Time : Fri Jan 27 15:55:12 2023 Raid Level : raid0 Array Size : 2549760 (2.43 GiB 2.61 GB) Raid Devices : 5 Total Devices : 5 Persistence : Superblock is persistent Update Time : Fri Jan 27 15:55:12 2023 State : clean Active Devices : 5 Working Devices : 5 Failed Devices : 0 Spare Devices : 0 Layout : -unknown- Chunk Size : 512K Consistency Policy : none Name : 1 UUID : 4820fee5:fd767c2b:e0973487:25f8de93 Events : 0 Number Major Minor RaidDevice State 0 7 0 0 active sync /dev/loop0 1 7 3 1 active sync /dev/loop3 2 7 1 2 active sync /dev/loop1 3 7 2 3 active sync /dev/loop2 4 7 4 4 active sync /dev/loop4 :: [ 15:55:20 ] :: [ PASS ] :: Command 'mdadm --detail /dev/md1' (Expected 0, got 0) :: [ 15:55:20 ] :: [ LOG ] :: INFO: Successfully created md raid /dev/md1 1674852920 start_time against this md array: /dev/md1 state is clean :: [ 15:55:20 ] :: [ LOG ] :: mkfs -t xfs /dev/md1 log stripe unit (524288 bytes) is too large (maximum is 256KiB) log stripe unit adjusted to 32KiB meta-data=/dev/md1 isize=512 agcount=8, agsize=79744 blks = sectsz=512 attr=2, projid32bit=1 = crc=1 finobt=1, sparse=1, rmapbt=0 = reflink=1 bigtime=1 inobtcount=1 data = bsize=4096 blocks=637440, imaxpct=25 = sunit=128 swidth=640 blks naming =version 2 bsize=4096 ascii-ci=0, ftype=1 log =internal log bsize=4096 blocks=2560, version=2 = sectsz=512 sunit=8 blks, lazy-count=1 realtime =none extsz=4096 blocks=0, rtextents=0 Discarding blocks...Done. :: [ 15:55:22 ] :: [ BEGIN ] :: Running 'mount -t xfs /dev/md1 /mnt/md_test ' :: [ 15:55:22 ] :: [ PASS ] :: Command 'mount -t xfs /dev/md1 /mnt/md_test ' (Expected 0, got 0) :: [ 15:55:22 ] :: [ BEGIN ] :: Running 'fstrim -v /mnt/md_test' /mnt/md_test: 2.4 GiB (2600108032 bytes) trimmed :: [ 15:55:23 ] :: [ PASS ] :: Command 'fstrim -v /mnt/md_test' (Expected 0, got 0) :: [ 15:55:23 ] :: [ BEGIN ] :: Running 'umount /dev/md1' :: [ 15:55:23 ] :: [ PASS ] :: Command 'umount /dev/md1' (Expected 0, got 0) INFO: Executing MD_Clean_RAID() against this md device: mdadm --stop /dev/md1 mdadm: stopped /dev/md1 clean devs : /dev/loop0 /dev/loop3 /dev/loop1 /dev/loop2 /dev/loop4 mdadm --zero-superblock /dev/loop0 mdadm --zero-superblock /dev/loop3 mdadm --zero-superblock /dev/loop1 mdadm --zero-superblock /dev/loop2 mdadm --zero-superblock /dev/loop4 ret is 0 ls /dev/md1 ls: cannot access '/dev/md1': No such file or directory mdadm --stop can delete md node name /dev/md1 in /dev INFO: Executing MD_Create_RAID() to create raid 1 INFO: Created md raid with these raid devices " /dev/loop0 /dev/loop3 /dev/loop1 /dev/loop2 /dev/loop4". INFO: Created md raid with these spare disks " /dev/loop5". :: [ 15:55:50 ] :: [ BEGIN ] :: Running 'mdadm --create --run /dev/md1 --level 1 --metadata 1.2 --raid-devices 5 /dev/loop0 /dev/loop3 /dev/loop1 /dev/loop2 /dev/loop4 --spare-devices 1 /dev/loop5 --bitmap=internal --bitmap-chunk=64M' mdadm: array /dev/md1 started. :: [ 15:55:50 ] :: [ PASS ] :: Command 'mdadm --create --run /dev/md1 --level 1 --metadata 1.2 --raid-devices 5 /dev/loop0 /dev/loop3 /dev/loop1 /dev/loop2 /dev/loop4 --spare-devices 1 /dev/loop5 --bitmap=internal --bitmap-chunk=64M' (Expected 0, got 0) INFO:cat /proc/mdstat###################### :: [ 15:55:51 ] :: [ BEGIN ] :: Running 'cat /proc/mdstat' Personalities : [raid6] [raid5] [raid4] [raid0] [raid1] md1 : active raid1 loop5[5](S) loop4[4] loop2[3] loop1[2] loop3[1] loop0[0] 510976 blocks super 1.2 [5/5] [UUUUU] [>....................] resync = 2.0% (10368/510976) finish=0.8min speed=10368K/sec bitmap: 1/1 pages [4KB], 65536KB chunk unused devices: :: [ 15:55:51 ] :: [ PASS ] :: Command 'cat /proc/mdstat' (Expected 0, got 0) :: [ 15:55:51 ] :: [ BEGIN ] :: Running 'lsblk' NAME MAJ:MIN RM SIZE RO TYPE MOUNTPOINTS loop0 7:0 0 500M 0 loop └─md1 9:1 0 499M 0 raid1 loop1 7:1 0 500M 0 loop └─md1 9:1 0 499M 0 raid1 loop2 7:2 0 500M 0 loop └─md1 9:1 0 499M 0 raid1 loop3 7:3 0 500M 0 loop └─md1 9:1 0 499M 0 raid1 loop4 7:4 0 500M 0 loop └─md1 9:1 0 499M 0 raid1 loop5 7:5 0 500M 0 loop └─md1 9:1 0 499M 0 raid1 sda 8:0 0 558.7G 0 disk ├─sda1 8:1 0 1G 0 part /boot └─sda2 8:2 0 557.7G 0 part ├─cs_hpe--ml350gen9--01-root 253:0 0 70G 0 lvm / ├─cs_hpe--ml350gen9--01-swap 253:1 0 23.6G 0 lvm [SWAP] └─cs_hpe--ml350gen9--01-home 253:2 0 464.2G 0 lvm /home :: [ 15:55:51 ] :: [ PASS ] :: Command 'lsblk' (Expected 0, got 0) /dev/md1 INFO:mdadm -D /dev/md1 ######################### :: [ 15:55:52 ] :: [ BEGIN ] :: Running 'mdadm --detail /dev/md1' /dev/md1: Version : 1.2 Creation Time : Fri Jan 27 15:55:50 2023 Raid Level : raid1 Array Size : 510976 (499.00 MiB 523.24 MB) Used Dev Size : 510976 (499.00 MiB 523.24 MB) Raid Devices : 5 Total Devices : 6 Persistence : Superblock is persistent Intent Bitmap : Internal Update Time : Fri Jan 27 15:55:51 2023 State : clean, resyncing Active Devices : 5 Working Devices : 6 Failed Devices : 0 Spare Devices : 1 Consistency Policy : bitmap Resync Status : 7% complete Name : 1 UUID : d54acd8f:81be5288:c8de3ac9:987e0fda Events : 1 Number Major Minor RaidDevice State 0 7 0 0 active sync /dev/loop0 1 7 3 1 active sync /dev/loop3 2 7 1 2 active sync /dev/loop1 3 7 2 3 active sync /dev/loop2 4 7 4 4 active sync /dev/loop4 5 7 5 - spare /dev/loop5 :: [ 15:55:52 ] :: [ PASS ] :: Command 'mdadm --detail /dev/md1' (Expected 0, got 0) :: [ 15:55:52 ] :: [ LOG ] :: INFO: Successfully created md raid /dev/md1 1674852952 start_time against this md array: /dev/md1 state is clean, 1674852958 start_time against this md array: /dev/md1 state is clean, 1674852963 start_time against this md array: /dev/md1 state is clean, 1674852969 start_time against this md array: /dev/md1 state is clean, 1674852974 start_time against this md array: /dev/md1 state is clean :: [ 15:56:14 ] :: [ LOG ] :: mkfs -t xfs /dev/md1 meta-data=/dev/md1 isize=512 agcount=4, agsize=31936 blks = sectsz=512 attr=2, projid32bit=1 = crc=1 finobt=1, sparse=1, rmapbt=0 = reflink=1 bigtime=1 inobtcount=1 data = bsize=4096 blocks=127744, imaxpct=25 = sunit=0 swidth=0 blks naming =version 2 bsize=4096 ascii-ci=0, ftype=1 log =internal log bsize=4096 blocks=1368, version=2 = sectsz=512 sunit=0 blks, lazy-count=1 realtime =none extsz=4096 blocks=0, rtextents=0 Discarding blocks...Done. :: [ 15:56:16 ] :: [ BEGIN ] :: Running 'mount -t xfs /dev/md1 /mnt/md_test ' :: [ 15:56:16 ] :: [ PASS ] :: Command 'mount -t xfs /dev/md1 /mnt/md_test ' (Expected 0, got 0) :: [ 15:56:17 ] :: [ BEGIN ] :: Running 'fstrim -v /mnt/md_test' /mnt/md_test: 493.5 MiB (517439488 bytes) trimmed :: [ 15:56:17 ] :: [ PASS ] :: Command 'fstrim -v /mnt/md_test' (Expected 0, got 0) :: [ 15:56:17 ] :: [ BEGIN ] :: Running 'umount /dev/md1' :: [ 15:56:18 ] :: [ PASS ] :: Command 'umount /dev/md1' (Expected 0, got 0) INFO: Executing MD_Clean_RAID() against this md device: mdadm --stop /dev/md1 mdadm: stopped /dev/md1 clean devs : /dev/loop0 /dev/loop3 /dev/loop1 /dev/loop2 /dev/loop4 /dev/loop5 mdadm --zero-superblock /dev/loop0 mdadm --zero-superblock /dev/loop3 mdadm --zero-superblock /dev/loop1 mdadm --zero-superblock /dev/loop2 mdadm --zero-superblock /dev/loop4 mdadm --zero-superblock /dev/loop5 ret is 0 ls /dev/md1 ls: cannot access '/dev/md1': No such file or directory mdadm --stop can delete md node name /dev/md1 in /dev INFO: Executing MD_Create_RAID() to create raid 4 INFO: Created md raid with these raid devices " /dev/loop0 /dev/loop3 /dev/loop1 /dev/loop2 /dev/loop4". INFO: Created md raid with these spare disks " /dev/loop5". :: [ 15:56:44 ] :: [ BEGIN ] :: Running 'mdadm --create --run /dev/md1 --level 4 --metadata 1.2 --raid-devices 5 /dev/loop0 /dev/loop3 /dev/loop1 /dev/loop2 /dev/loop4 --spare-devices 1 /dev/loop5 --chunk 512 --bitmap=internal --bitmap-chunk=64M' mdadm: array /dev/md1 started. :: [ 15:56:45 ] :: [ PASS ] :: Command 'mdadm --create --run /dev/md1 --level 4 --metadata 1.2 --raid-devices 5 /dev/loop0 /dev/loop3 /dev/loop1 /dev/loop2 /dev/loop4 --spare-devices 1 /dev/loop5 --chunk 512 --bitmap=internal --bitmap-chunk=64M' (Expected 0, got 0) INFO:cat /proc/mdstat###################### :: [ 15:56:45 ] :: [ BEGIN ] :: Running 'cat /proc/mdstat' Personalities : [raid6] [raid5] [raid4] [raid0] [raid1] md1 : active raid4 loop4[6] loop5[5](S) loop2[3] loop1[2] loop3[1] loop0[0] 2039808 blocks super 1.2 level 4, 512k chunk, algorithm 0 [5/4] [UUUU_] [>....................] recovery = 0.6% (3280/509952) finish=2.5min speed=3280K/sec bitmap: 1/1 pages [4KB], 65536KB chunk unused devices: :: [ 15:56:45 ] :: [ PASS ] :: Command 'cat /proc/mdstat' (Expected 0, got 0) :: [ 15:56:45 ] :: [ BEGIN ] :: Running 'lsblk' NAME MAJ:MIN RM SIZE RO TYPE MOUNTPOINTS loop0 7:0 0 500M 0 loop └─md1 9:1 0 1.9G 0 raid4 loop1 7:1 0 500M 0 loop └─md1 9:1 0 1.9G 0 raid4 loop2 7:2 0 500M 0 loop └─md1 9:1 0 1.9G 0 raid4 loop3 7:3 0 500M 0 loop └─md1 9:1 0 1.9G 0 raid4 loop4 7:4 0 500M 0 loop └─md1 9:1 0 1.9G 0 raid4 loop5 7:5 0 500M 0 loop └─md1 9:1 0 1.9G 0 raid4 sda 8:0 0 558.7G 0 disk ├─sda1 8:1 0 1G 0 part /boot └─sda2 8:2 0 557.7G 0 part ├─cs_hpe--ml350gen9--01-root 253:0 0 70G 0 lvm / ├─cs_hpe--ml350gen9--01-swap 253:1 0 23.6G 0 lvm [SWAP] └─cs_hpe--ml350gen9--01-home 253:2 0 464.2G 0 lvm /home :: [ 15:56:46 ] :: [ PASS ] :: Command 'lsblk' (Expected 0, got 0) /dev/md1 INFO:mdadm -D /dev/md1 ######################### :: [ 15:56:46 ] :: [ BEGIN ] :: Running 'mdadm --detail /dev/md1' /dev/md1: Version : 1.2 Creation Time : Fri Jan 27 15:56:44 2023 Raid Level : raid4 Array Size : 2039808 (1992.00 MiB 2088.76 MB) Used Dev Size : 509952 (498.00 MiB 522.19 MB) Raid Devices : 5 Total Devices : 6 Persistence : Superblock is persistent Intent Bitmap : Internal Update Time : Fri Jan 27 15:56:44 2023 State : clean, degraded, recovering Active Devices : 4 Working Devices : 6 Failed Devices : 0 Spare Devices : 2 Chunk Size : 512K Consistency Policy : bitmap Rebuild Status : 4% complete Name : 1 UUID : 72177688:93527198:4668a6bc:29ff5a86 Events : 1 Number Major Minor RaidDevice State 0 7 0 0 active sync /dev/loop0 1 7 3 1 active sync /dev/loop3 2 7 1 2 active sync /dev/loop1 3 7 2 3 active sync /dev/loop2 6 7 4 4 spare rebuilding /dev/loop4 5 7 5 - spare /dev/loop5 :: [ 15:56:46 ] :: [ PASS ] :: Command 'mdadm --detail /dev/md1' (Expected 0, got 0) :: [ 15:56:46 ] :: [ LOG ] :: INFO: Successfully created md raid /dev/md1 1674853007 start_time against this md array: /dev/md1 state is clean, 1674853012 start_time against this md array: /dev/md1 state is clean, 1674853017 start_time against this md array: /dev/md1 state is clean, 1674853022 start_time against this md array: /dev/md1 state is clean, 1674853028 start_time against this md array: /dev/md1 state is clean, 1674853033 start_time against this md array: /dev/md1 state is clean, 1674853038 start_time against this md array: /dev/md1 state is clean, 1674853043 start_time against this md array: /dev/md1 state is clean :: [ 15:57:23 ] :: [ LOG ] :: mkfs -t xfs /dev/md1 log stripe unit (524288 bytes) is too large (maximum is 256KiB) log stripe unit adjusted to 32KiB meta-data=/dev/md1 isize=512 agcount=8, agsize=63744 blks = sectsz=512 attr=2, projid32bit=1 = crc=1 finobt=1, sparse=1, rmapbt=0 = reflink=1 bigtime=1 inobtcount=1 data = bsize=4096 blocks=509952, imaxpct=25 = sunit=128 swidth=512 blks naming =version 2 bsize=4096 ascii-ci=0, ftype=1 log =internal log bsize=4096 blocks=2560, version=2 = sectsz=512 sunit=8 blks, lazy-count=1 realtime =none extsz=4096 blocks=0, rtextents=0 Discarding blocks...Done. :: [ 15:58:33 ] :: [ BEGIN ] :: Running 'mount -t xfs /dev/md1 /mnt/md_test ' :: [ 15:58:34 ] :: [ PASS ] :: Command 'mount -t xfs /dev/md1 /mnt/md_test ' (Expected 0, got 0) :: [ 15:58:34 ] :: [ BEGIN ] :: Running 'fstrim -v /mnt/md_test' /mnt/md_test: 1.9 GiB (2077425664 bytes) trimmed :: [ 15:59:03 ] :: [ PASS ] :: Command 'fstrim -v /mnt/md_test' (Expected 0, got 0) :: [ 15:59:03 ] :: [ BEGIN ] :: Running 'umount /dev/md1' :: [ 15:59:03 ] :: [ PASS ] :: Command 'umount /dev/md1' (Expected 0, got 0) INFO: Executing MD_Clean_RAID() against this md device: mdadm --stop /dev/md1 mdadm: stopped /dev/md1 clean devs : /dev/loop0 /dev/loop3 /dev/loop1 /dev/loop2 /dev/loop4 /dev/loop5 mdadm --zero-superblock /dev/loop0 mdadm --zero-superblock /dev/loop3 mdadm --zero-superblock /dev/loop1 mdadm --zero-superblock /dev/loop2 mdadm --zero-superblock /dev/loop4 mdadm --zero-superblock /dev/loop5 ret is 0 ls /dev/md1 ls: cannot access '/dev/md1': No such file or directory mdadm --stop can delete md node name /dev/md1 in /dev INFO: Executing MD_Create_RAID() to create raid 5 INFO: Created md raid with these raid devices " /dev/loop0 /dev/loop3 /dev/loop1 /dev/loop2 /dev/loop4". INFO: Created md raid with these spare disks " /dev/loop5". :: [ 15:59:30 ] :: [ BEGIN ] :: Running 'mdadm --create --run /dev/md1 --level 5 --metadata 1.2 --raid-devices 5 /dev/loop0 /dev/loop3 /dev/loop1 /dev/loop2 /dev/loop4 --spare-devices 1 /dev/loop5 --chunk 512 --bitmap=internal --bitmap-chunk=64M' mdadm: array /dev/md1 started. :: [ 15:59:30 ] :: [ PASS ] :: Command 'mdadm --create --run /dev/md1 --level 5 --metadata 1.2 --raid-devices 5 /dev/loop0 /dev/loop3 /dev/loop1 /dev/loop2 /dev/loop4 --spare-devices 1 /dev/loop5 --chunk 512 --bitmap=internal --bitmap-chunk=64M' (Expected 0, got 0) INFO:cat /proc/mdstat###################### :: [ 15:59:31 ] :: [ BEGIN ] :: Running 'cat /proc/mdstat' Personalities : [raid6] [raid5] [raid4] [raid0] [raid1] md1 : active raid5 loop4[6] loop5[5](S) loop2[3] loop1[2] loop3[1] loop0[0] 2039808 blocks super 1.2 level 5, 512k chunk, algorithm 2 [5/4] [UUUU_] [>....................] recovery = 1.0% (5508/509952) finish=1.5min speed=5508K/sec bitmap: 1/1 pages [4KB], 65536KB chunk unused devices: :: [ 15:59:31 ] :: [ PASS ] :: Command 'cat /proc/mdstat' (Expected 0, got 0) :: [ 15:59:31 ] :: [ BEGIN ] :: Running 'lsblk' NAME MAJ:MIN RM SIZE RO TYPE MOUNTPOINTS loop0 7:0 0 500M 0 loop └─md1 9:1 0 1.9G 0 raid5 loop1 7:1 0 500M 0 loop └─md1 9:1 0 1.9G 0 raid5 loop2 7:2 0 500M 0 loop └─md1 9:1 0 1.9G 0 raid5 loop3 7:3 0 500M 0 loop └─md1 9:1 0 1.9G 0 raid5 loop4 7:4 0 500M 0 loop └─md1 9:1 0 1.9G 0 raid5 loop5 7:5 0 500M 0 loop └─md1 9:1 0 1.9G 0 raid5 sda 8:0 0 558.7G 0 disk ├─sda1 8:1 0 1G 0 part /boot └─sda2 8:2 0 557.7G 0 part ├─cs_hpe--ml350gen9--01-root 253:0 0 70G 0 lvm / ├─cs_hpe--ml350gen9--01-swap 253:1 0 23.6G 0 lvm [SWAP] └─cs_hpe--ml350gen9--01-home 253:2 0 464.2G 0 lvm /home :: [ 15:59:31 ] :: [ PASS ] :: Command 'lsblk' (Expected 0, got 0) /dev/md1 INFO:mdadm -D /dev/md1 ######################### :: [ 15:59:32 ] :: [ BEGIN ] :: Running 'mdadm --detail /dev/md1' /dev/md1: Version : 1.2 Creation Time : Fri Jan 27 15:59:30 2023 Raid Level : raid5 Array Size : 2039808 (1992.00 MiB 2088.76 MB) Used Dev Size : 509952 (498.00 MiB 522.19 MB) Raid Devices : 5 Total Devices : 6 Persistence : Superblock is persistent Intent Bitmap : Internal Update Time : Fri Jan 27 15:59:30 2023 State : clean, degraded, recovering Active Devices : 4 Working Devices : 6 Failed Devices : 0 Spare Devices : 2 Layout : left-symmetric Chunk Size : 512K Consistency Policy : bitmap Rebuild Status : 4% complete Name : 1 UUID : 935fe7c0:e4e50790:3d0fb13c:b59b0729 Events : 1 Number Major Minor RaidDevice State 0 7 0 0 active sync /dev/loop0 1 7 3 1 active sync /dev/loop3 2 7 1 2 active sync /dev/loop1 3 7 2 3 active sync /dev/loop2 6 7 4 4 spare rebuilding /dev/loop4 5 7 5 - spare /dev/loop5 :: [ 15:59:32 ] :: [ PASS ] :: Command 'mdadm --detail /dev/md1' (Expected 0, got 0) :: [ 15:59:32 ] :: [ LOG ] :: INFO: Successfully created md raid /dev/md1 1674853172 start_time against this md array: /dev/md1 state is clean, 1674853178 start_time against this md array: /dev/md1 state is clean, 1674853183 start_time against this md array: /dev/md1 state is clean, 1674853188 start_time against this md array: /dev/md1 state is clean, 1674853193 start_time against this md array: /dev/md1 state is clean, 1674853199 start_time against this md array: /dev/md1 state is clean, 1674853204 start_time against this md array: /dev/md1 state is clean, 1674853209 start_time against this md array: /dev/md1 state is clean :: [ 16:00:09 ] :: [ LOG ] :: mkfs -t xfs /dev/md1 mkfs.xfs: /dev/md1 appears to contain an existing filesystem (xfs). mkfs.xfs: Use the -f option to force overwrite. log stripe unit (524288 bytes) is too large (maximum is 256KiB) log stripe unit adjusted to 32KiB meta-data=/dev/md1 isize=512 agcount=8, agsize=63744 blks = sectsz=512 attr=2, projid32bit=1 = crc=1 finobt=1, sparse=1, rmapbt=0 = reflink=1 bigtime=1 inobtcount=1 data = bsize=4096 blocks=509952, imaxpct=25 = sunit=128 swidth=512 blks naming =version 2 bsize=4096 ascii-ci=0, ftype=1 log =internal log bsize=4096 blocks=2560, version=2 = sectsz=512 sunit=8 blks, lazy-count=1 realtime =none extsz=4096 blocks=0, rtextents=0 Discarding blocks...Done. :: [ 16:01:19 ] :: [ BEGIN ] :: Running 'mount -t xfs /dev/md1 /mnt/md_test ' :: [ 16:01:19 ] :: [ PASS ] :: Command 'mount -t xfs /dev/md1 /mnt/md_test ' (Expected 0, got 0) :: [ 16:01:19 ] :: [ BEGIN ] :: Running 'fstrim -v /mnt/md_test' /mnt/md_test: 1.9 GiB (2077425664 bytes) trimmed :: [ 16:01:49 ] :: [ PASS ] :: Command 'fstrim -v /mnt/md_test' (Expected 0, got 0) :: [ 16:01:50 ] :: [ BEGIN ] :: Running 'umount /dev/md1' :: [ 16:01:50 ] :: [ PASS ] :: Command 'umount /dev/md1' (Expected 0, got 0) INFO: Executing MD_Clean_RAID() against this md device: mdadm --stop /dev/md1 mdadm: stopped /dev/md1 clean devs : /dev/loop0 /dev/loop3 /dev/loop1 /dev/loop2 /dev/loop4 /dev/loop5 mdadm --zero-superblock /dev/loop0 mdadm --zero-superblock /dev/loop3 mdadm --zero-superblock /dev/loop1 mdadm --zero-superblock /dev/loop2 mdadm --zero-superblock /dev/loop4 mdadm --zero-superblock /dev/loop5 ret is 0 ls /dev/md1 ls: cannot access '/dev/md1': No such file or directory mdadm --stop can delete md node name /dev/md1 in /dev INFO: Executing MD_Create_RAID() to create raid 6 INFO: Created md raid with these raid devices " /dev/loop0 /dev/loop3 /dev/loop1 /dev/loop2 /dev/loop4". INFO: Created md raid with these spare disks " /dev/loop5". :: [ 16:02:17 ] :: [ BEGIN ] :: Running 'mdadm --create --run /dev/md1 --level 6 --metadata 1.2 --raid-devices 5 /dev/loop0 /dev/loop3 /dev/loop1 /dev/loop2 /dev/loop4 --spare-devices 1 /dev/loop5 --chunk 512 --bitmap=internal --bitmap-chunk=64M' mdadm: array /dev/md1 started. :: [ 16:02:17 ] :: [ PASS ] :: Command 'mdadm --create --run /dev/md1 --level 6 --metadata 1.2 --raid-devices 5 /dev/loop0 /dev/loop3 /dev/loop1 /dev/loop2 /dev/loop4 --spare-devices 1 /dev/loop5 --chunk 512 --bitmap=internal --bitmap-chunk=64M' (Expected 0, got 0) INFO:cat /proc/mdstat###################### :: [ 16:02:18 ] :: [ BEGIN ] :: Running 'cat /proc/mdstat' Personalities : [raid6] [raid5] [raid4] [raid0] [raid1] md1 : active raid6 loop5[5](S) loop4[4] loop2[3] loop1[2] loop3[1] loop0[0] 1529856 blocks super 1.2 level 6, 512k chunk, algorithm 2 [5/5] [UUUUU] [>....................] resync = 1.5% (8004/509952) finish=1.0min speed=8004K/sec bitmap: 1/1 pages [4KB], 65536KB chunk unused devices: :: [ 16:02:18 ] :: [ PASS ] :: Command 'cat /proc/mdstat' (Expected 0, got 0) :: [ 16:02:18 ] :: [ BEGIN ] :: Running 'lsblk' NAME MAJ:MIN RM SIZE RO TYPE MOUNTPOINTS loop0 7:0 0 500M 0 loop └─md1 9:1 0 1.5G 0 raid6 loop1 7:1 0 500M 0 loop └─md1 9:1 0 1.5G 0 raid6 loop2 7:2 0 500M 0 loop └─md1 9:1 0 1.5G 0 raid6 loop3 7:3 0 500M 0 loop └─md1 9:1 0 1.5G 0 raid6 loop4 7:4 0 500M 0 loop └─md1 9:1 0 1.5G 0 raid6 loop5 7:5 0 500M 0 loop └─md1 9:1 0 1.5G 0 raid6 sda 8:0 0 558.7G 0 disk ├─sda1 8:1 0 1G 0 part /boot └─sda2 8:2 0 557.7G 0 part ├─cs_hpe--ml350gen9--01-root 253:0 0 70G 0 lvm / ├─cs_hpe--ml350gen9--01-swap 253:1 0 23.6G 0 lvm [SWAP] └─cs_hpe--ml350gen9--01-home 253:2 0 464.2G 0 lvm /home :: [ 16:02:18 ] :: [ PASS ] :: Command 'lsblk' (Expected 0, got 0) /dev/md1 INFO:mdadm -D /dev/md1 ######################### :: [ 16:02:19 ] :: [ BEGIN ] :: Running 'mdadm --detail /dev/md1' /dev/md1: Version : 1.2 Creation Time : Fri Jan 27 16:02:17 2023 Raid Level : raid6 Array Size : 1529856 (1494.00 MiB 1566.57 MB) Used Dev Size : 509952 (498.00 MiB 522.19 MB) Raid Devices : 5 Total Devices : 6 Persistence : Superblock is persistent Intent Bitmap : Internal Update Time : Fri Jan 27 16:02:17 2023 State : clean, resyncing Active Devices : 5 Working Devices : 6 Failed Devices : 0 Spare Devices : 1 Layout : left-symmetric Chunk Size : 512K Consistency Policy : bitmap Resync Status : 6% complete Name : 1 UUID : f8a68366:3cc1d621:ec2b51c1:7b086389 Events : 0 Number Major Minor RaidDevice State 0 7 0 0 active sync /dev/loop0 1 7 3 1 active sync /dev/loop3 2 7 1 2 active sync /dev/loop1 3 7 2 3 active sync /dev/loop2 4 7 4 4 active sync /dev/loop4 5 7 5 - spare /dev/loop5 :: [ 16:02:19 ] :: [ PASS ] :: Command 'mdadm --detail /dev/md1' (Expected 0, got 0) :: [ 16:02:19 ] :: [ LOG ] :: INFO: Successfully created md raid /dev/md1 1674853339 start_time against this md array: /dev/md1 state is clean, 1674853345 start_time against this md array: /dev/md1 state is clean, 1674853350 start_time against this md array: /dev/md1 state is clean, 1674853355 start_time against this md array: /dev/md1 state is clean, 1674853360 start_time against this md array: /dev/md1 state is clean, 1674853365 start_time against this md array: /dev/md1 state is clean :: [ 16:02:46 ] :: [ LOG ] :: mkfs -t xfs /dev/md1 log stripe unit (524288 bytes) is too large (maximum is 256KiB) log stripe unit adjusted to 32KiB meta-data=/dev/md1 isize=512 agcount=8, agsize=47872 blks = sectsz=512 attr=2, projid32bit=1 = crc=1 finobt=1, sparse=1, rmapbt=0 = reflink=1 bigtime=1 inobtcount=1 data = bsize=4096 blocks=382464, imaxpct=25 = sunit=128 swidth=384 blks naming =version 2 bsize=4096 ascii-ci=0, ftype=1 log =internal log bsize=4096 blocks=2560, version=2 = sectsz=512 sunit=8 blks, lazy-count=1 realtime =none extsz=4096 blocks=0, rtextents=0 Discarding blocks...Done. :: [ 16:03:20 ] :: [ BEGIN ] :: Running 'mount -t xfs /dev/md1 /mnt/md_test ' :: [ 16:03:21 ] :: [ PASS ] :: Command 'mount -t xfs /dev/md1 /mnt/md_test ' (Expected 0, got 0) :: [ 16:03:21 ] :: [ BEGIN ] :: Running 'fstrim -v /mnt/md_test' /mnt/md_test: 1.4 GiB (1555234816 bytes) trimmed :: [ 16:03:46 ] :: [ PASS ] :: Command 'fstrim -v /mnt/md_test' (Expected 0, got 0) :: [ 16:03:47 ] :: [ BEGIN ] :: Running 'umount /dev/md1' :: [ 16:03:47 ] :: [ PASS ] :: Command 'umount /dev/md1' (Expected 0, got 0) INFO: Executing MD_Clean_RAID() against this md device: mdadm --stop /dev/md1 mdadm: stopped /dev/md1 clean devs : /dev/loop0 /dev/loop3 /dev/loop1 /dev/loop2 /dev/loop4 /dev/loop5 mdadm --zero-superblock /dev/loop0 mdadm --zero-superblock /dev/loop3 mdadm --zero-superblock /dev/loop1 mdadm --zero-superblock /dev/loop2 mdadm --zero-superblock /dev/loop4 mdadm --zero-superblock /dev/loop5 ret is 0 ls /dev/md1 ls: cannot access '/dev/md1': No such file or directory mdadm --stop can delete md node name /dev/md1 in /dev INFO: Executing MD_Create_RAID() to create raid 10 INFO: Created md raid with these raid devices " /dev/loop0 /dev/loop3 /dev/loop1 /dev/loop2 /dev/loop4". INFO: Created md raid with these spare disks " /dev/loop5". :: [ 16:04:14 ] :: [ BEGIN ] :: Running 'mdadm --create --run /dev/md1 --level 10 --metadata 1.2 --raid-devices 5 /dev/loop0 /dev/loop3 /dev/loop1 /dev/loop2 /dev/loop4 --spare-devices 1 /dev/loop5 --chunk 512 --bitmap=internal --bitmap-chunk=64M' mdadm: array /dev/md1 started. :: [ 16:04:14 ] :: [ PASS ] :: Command 'mdadm --create --run /dev/md1 --level 10 --metadata 1.2 --raid-devices 5 /dev/loop0 /dev/loop3 /dev/loop1 /dev/loop2 /dev/loop4 --spare-devices 1 /dev/loop5 --chunk 512 --bitmap=internal --bitmap-chunk=64M' (Expected 0, got 0) INFO:cat /proc/mdstat###################### :: [ 16:04:15 ] :: [ BEGIN ] :: Running 'cat /proc/mdstat' Personalities : [raid6] [raid5] [raid4] [raid0] [raid1] [raid10] md1 : active raid10 loop5[5](S) loop4[4] loop2[3] loop1[2] loop3[1] loop0[0] 1274880 blocks super 1.2 512K chunks 2 near-copies [5/5] [UUUUU] [>....................] resync = 3.4% (44416/1274880) finish=0.4min speed=44416K/sec bitmap: 1/1 pages [4KB], 65536KB chunk unused devices: :: [ 16:04:15 ] :: [ PASS ] :: Command 'cat /proc/mdstat' (Expected 0, got 0) :: [ 16:04:15 ] :: [ BEGIN ] :: Running 'lsblk' NAME MAJ:MIN RM SIZE RO TYPE MOUNTPOINTS loop0 7:0 0 500M 0 loop └─md1 9:1 0 1.2G 0 raid10 loop1 7:1 0 500M 0 loop └─md1 9:1 0 1.2G 0 raid10 loop2 7:2 0 500M 0 loop └─md1 9:1 0 1.2G 0 raid10 loop3 7:3 0 500M 0 loop └─md1 9:1 0 1.2G 0 raid10 loop4 7:4 0 500M 0 loop └─md1 9:1 0 1.2G 0 raid10 loop5 7:5 0 500M 0 loop └─md1 9:1 0 1.2G 0 raid10 sda 8:0 0 558.7G 0 disk ├─sda1 8:1 0 1G 0 part /boot └─sda2 8:2 0 557.7G 0 part ├─cs_hpe--ml350gen9--01-root 253:0 0 70G 0 lvm / ├─cs_hpe--ml350gen9--01-swap 253:1 0 23.6G 0 lvm [SWAP] └─cs_hpe--ml350gen9--01-home 253:2 0 464.2G 0 lvm /home :: [ 16:04:15 ] :: [ PASS ] :: Command 'lsblk' (Expected 0, got 0) /dev/md1 INFO:mdadm -D /dev/md1 ######################### :: [ 16:04:16 ] :: [ BEGIN ] :: Running 'mdadm --detail /dev/md1' /dev/md1: Version : 1.2 Creation Time : Fri Jan 27 16:04:14 2023 Raid Level : raid10 Array Size : 1274880 (1245.00 MiB 1305.48 MB) Used Dev Size : 509952 (498.00 MiB 522.19 MB) Raid Devices : 5 Total Devices : 6 Persistence : Superblock is persistent Intent Bitmap : Internal Update Time : Fri Jan 27 16:04:15 2023 State : clean, resyncing Active Devices : 5 Working Devices : 6 Failed Devices : 0 Spare Devices : 1 Layout : near=2 Chunk Size : 512K Consistency Policy : bitmap Resync Status : 16% complete Name : 1 UUID : d6200d7f:bae3b67d:88e80d64:14aba3cb Events : 2 Number Major Minor RaidDevice State 0 7 0 0 active sync /dev/loop0 1 7 3 1 active sync /dev/loop3 2 7 1 2 active sync /dev/loop1 3 7 2 3 active sync /dev/loop2 4 7 4 4 active sync /dev/loop4 5 7 5 - spare /dev/loop5 :: [ 16:04:16 ] :: [ PASS ] :: Command 'mdadm --detail /dev/md1' (Expected 0, got 0) :: [ 16:04:16 ] :: [ LOG ] :: INFO: Successfully created md raid /dev/md1 1674853456 start_time against this md array: /dev/md1 state is clean, 1674853462 start_time against this md array: /dev/md1 state is clean, 1674853467 start_time against this md array: /dev/md1 state is clean :: [ 16:04:27 ] :: [ LOG ] :: mkfs -t xfs /dev/md1 log stripe unit (524288 bytes) is too large (maximum is 256KiB) log stripe unit adjusted to 32KiB meta-data=/dev/md1 isize=512 agcount=8, agsize=39936 blks = sectsz=512 attr=2, projid32bit=1 = crc=1 finobt=1, sparse=1, rmapbt=0 = reflink=1 bigtime=1 inobtcount=1 data = bsize=4096 blocks=318720, imaxpct=25 = sunit=128 swidth=640 blks naming =version 2 bsize=4096 ascii-ci=0, ftype=1 log =internal log bsize=4096 blocks=2560, version=2 = sectsz=512 sunit=8 blks, lazy-count=1 realtime =none extsz=4096 blocks=0, rtextents=0 Discarding blocks...Done. :: [ 16:04:29 ] :: [ BEGIN ] :: Running 'mount -t xfs /dev/md1 /mnt/md_test ' :: [ 16:04:29 ] :: [ PASS ] :: Command 'mount -t xfs /dev/md1 /mnt/md_test ' (Expected 0, got 0) :: [ 16:04:29 ] :: [ BEGIN ] :: Running 'fstrim -v /mnt/md_test' /mnt/md_test: 1.2 GiB (1294630912 bytes) trimmed :: [ 16:04:30 ] :: [ PASS ] :: Command 'fstrim -v /mnt/md_test' (Expected 0, got 0) :: [ 16:04:30 ] :: [ BEGIN ] :: Running 'umount /dev/md1' :: [ 16:04:31 ] :: [ PASS ] :: Command 'umount /dev/md1' (Expected 0, got 0) INFO: Executing MD_Clean_RAID() against this md device: mdadm --stop /dev/md1 mdadm: stopped /dev/md1 clean devs : /dev/loop0 /dev/loop3 /dev/loop1 /dev/loop2 /dev/loop4 /dev/loop5 mdadm --zero-superblock /dev/loop0 mdadm --zero-superblock /dev/loop3 mdadm --zero-superblock /dev/loop1 mdadm --zero-superblock /dev/loop2 mdadm --zero-superblock /dev/loop4 mdadm --zero-superblock /dev/loop5 ret is 0 ls /dev/md1 ls: cannot access '/dev/md1': No such file or directory mdadm --stop can delete md node name /dev/md1 in /dev losetup -d /dev/loop0 losetup -d /dev/loop3 losetup -d /dev/loop1 losetup -d /dev/loop2 losetup -d /dev/loop4 losetup -d /dev/loop5 :: [ 16:04:52 ] :: [ BEGIN ] :: check the errors :: actually running 'dmesg | grep -i 'Call Trace:'' :: [ 16:04:53 ] :: [ PASS ] :: check the errors (Expected 1, got 1) :::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::: :: Duration: 605s :: Assertions: 47 good, 0 bad :: RESULT: PASS (Test) ** Test PASS Score:0 Uploading resultoutputfile.log .done :::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::: :: TEST PROTOCOL :::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::: Test run ID : 9 Package : unknown beakerlib RPM : beakerlib-1.29.2-2.el9.noarch bl-redhat RPM : beakerlib-redhat-1-33.el9.noarch Test name : unknown Test started : 2023-01-27 15:54:45 EST Test finished : 2023-01-27 16:04:58 EST (still running) Test duration : 613 seconds Distro : CentOS Stream release 9 Hostname : hpe-ml350gen9-01.hpe2.lab.eng.bos.redhat.com Architecture : x86_64 CPUs : 72 x Intel(R) Xeon(R) CPU E5-2699 v3 @ 2.30GHz RAM size : 45650 MB HDD size : 534.87 GB :::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::: :: Test description :::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::: PURPOSE of trim Description: Test the function of TRIM which quickly erase invalid data for the RAID that created by mdadm. :::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::: :: Test :::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::: :: [ 15:54:48 ] :: [ PASS ] :: Command 'uname -a' (Expected 0, got 0) :: [ 15:54:50 ] :: [ PASS ] :: Command 'rpm -q mdadm || dnf install -y mdadm' (Expected 0, got 0) :: [ 15:54:50 ] :: [ LOG ] :: ./runtest.sh :: [ 15:54:50 ] :: [ PASS ] :: Command 'modprobe raid456 devices_handle_discard_safely=Y' (Expected 0, got 0) :: [ 15:54:51 ] :: [ PASS ] :: Command 'echo Y >/sys/module/raid456/parameters/devices_handle_discard_safely' (Expected 0, got 0) :: [ 15:55:18 ] :: [ PASS ] :: Command 'mdadm --create --run /dev/md1 --level 0 --metadata 1.2 --raid-devices 5 /dev/loop0 /dev/loop3 /dev/loop1 /dev/loop2 /dev/loop4 --chunk 512' (Expected 0, got 0) :: [ 15:55:19 ] :: [ PASS ] :: Command 'cat /proc/mdstat' (Expected 0, got 0) :: [ 15:55:19 ] :: [ PASS ] :: Command 'lsblk' (Expected 0, got 0) :: [ 15:55:20 ] :: [ PASS ] :: Command 'mdadm --detail /dev/md1' (Expected 0, got 0) :: [ 15:55:20 ] :: [ LOG ] :: INFO: Successfully created md raid /dev/md1 :: [ 15:55:20 ] :: [ LOG ] :: mkfs -t xfs /dev/md1 :: [ 15:55:22 ] :: [ PASS ] :: Command 'mount -t xfs /dev/md1 /mnt/md_test ' (Expected 0, got 0) :: [ 15:55:23 ] :: [ PASS ] :: Command 'fstrim -v /mnt/md_test' (Expected 0, got 0) :: [ 15:55:23 ] :: [ PASS ] :: Command 'umount /dev/md1' (Expected 0, got 0) :: [ 15:55:50 ] :: [ PASS ] :: Command 'mdadm --create --run /dev/md1 --level 1 --metadata 1.2 --raid-devices 5 /dev/loop0 /dev/loop3 /dev/loop1 /dev/loop2 /dev/loop4 --spare-devices 1 /dev/loop5 --bitmap=internal --bitmap-chunk=64M' (Expected 0, got 0) :: [ 15:55:51 ] :: [ PASS ] :: Command 'cat /proc/mdstat' (Expected 0, got 0) :: [ 15:55:51 ] :: [ PASS ] :: Command 'lsblk' (Expected 0, got 0) :: [ 15:55:52 ] :: [ PASS ] :: Command 'mdadm --detail /dev/md1' (Expected 0, got 0) :: [ 15:55:52 ] :: [ LOG ] :: INFO: Successfully created md raid /dev/md1 :: [ 15:56:14 ] :: [ LOG ] :: mkfs -t xfs /dev/md1 :: [ 15:56:16 ] :: [ PASS ] :: Command 'mount -t xfs /dev/md1 /mnt/md_test ' (Expected 0, got 0) :: [ 15:56:17 ] :: [ PASS ] :: Command 'fstrim -v /mnt/md_test' (Expected 0, got 0) :: [ 15:56:18 ] :: [ PASS ] :: Command 'umount /dev/md1' (Expected 0, got 0) :: [ 15:56:45 ] :: [ PASS ] :: Command 'mdadm --create --run /dev/md1 --level 4 --metadata 1.2 --raid-devices 5 /dev/loop0 /dev/loop3 /dev/loop1 /dev/loop2 /dev/loop4 --spare-devices 1 /dev/loop5 --chunk 512 --bitmap=internal --bitmap-chunk=64M' (Expected 0, got 0) :: [ 15:56:45 ] :: [ PASS ] :: Command 'cat /proc/mdstat' (Expected 0, got 0) :: [ 15:56:46 ] :: [ PASS ] :: Command 'lsblk' (Expected 0, got 0) :: [ 15:56:46 ] :: [ PASS ] :: Command 'mdadm --detail /dev/md1' (Expected 0, got 0) :: [ 15:56:46 ] :: [ LOG ] :: INFO: Successfully created md raid /dev/md1 :: [ 15:57:23 ] :: [ LOG ] :: mkfs -t xfs /dev/md1 :: [ 15:58:34 ] :: [ PASS ] :: Command 'mount -t xfs /dev/md1 /mnt/md_test ' (Expected 0, got 0) :: [ 15:59:03 ] :: [ PASS ] :: Command 'fstrim -v /mnt/md_test' (Expected 0, got 0) :: [ 15:59:03 ] :: [ PASS ] :: Command 'umount /dev/md1' (Expected 0, got 0) :: [ 15:59:30 ] :: [ PASS ] :: Command 'mdadm --create --run /dev/md1 --level 5 --metadata 1.2 --raid-devices 5 /dev/loop0 /dev/loop3 /dev/loop1 /dev/loop2 /dev/loop4 --spare-devices 1 /dev/loop5 --chunk 512 --bitmap=internal --bitmap-chunk=64M' (Expected 0, got 0) :: [ 15:59:31 ] :: [ PASS ] :: Command 'cat /proc/mdstat' (Expected 0, got 0) :: [ 15:59:31 ] :: [ PASS ] :: Command 'lsblk' (Expected 0, got 0) :: [ 15:59:32 ] :: [ PASS ] :: Command 'mdadm --detail /dev/md1' (Expected 0, got 0) :: [ 15:59:32 ] :: [ LOG ] :: INFO: Successfully created md raid /dev/md1 :: [ 16:00:09 ] :: [ LOG ] :: mkfs -t xfs /dev/md1 :: [ 16:01:19 ] :: [ PASS ] :: Command 'mount -t xfs /dev/md1 /mnt/md_test ' (Expected 0, got 0) :: [ 16:01:49 ] :: [ PASS ] :: Command 'fstrim -v /mnt/md_test' (Expected 0, got 0) :: [ 16:01:50 ] :: [ PASS ] :: Command 'umount /dev/md1' (Expected 0, got 0) :: [ 16:02:17 ] :: [ PASS ] :: Command 'mdadm --create --run /dev/md1 --level 6 --metadata 1.2 --raid-devices 5 /dev/loop0 /dev/loop3 /dev/loop1 /dev/loop2 /dev/loop4 --spare-devices 1 /dev/loop5 --chunk 512 --bitmap=internal --bitmap-chunk=64M' (Expected 0, got 0) :: [ 16:02:18 ] :: [ PASS ] :: Command 'cat /proc/mdstat' (Expected 0, got 0) :: [ 16:02:18 ] :: [ PASS ] :: Command 'lsblk' (Expected 0, got 0) :: [ 16:02:19 ] :: [ PASS ] :: Command 'mdadm --detail /dev/md1' (Expected 0, got 0) :: [ 16:02:19 ] :: [ LOG ] :: INFO: Successfully created md raid /dev/md1 :: [ 16:02:46 ] :: [ LOG ] :: mkfs -t xfs /dev/md1 :: [ 16:03:21 ] :: [ PASS ] :: Command 'mount -t xfs /dev/md1 /mnt/md_test ' (Expected 0, got 0) :: [ 16:03:46 ] :: [ PASS ] :: Command 'fstrim -v /mnt/md_test' (Expected 0, got 0) :: [ 16:03:47 ] :: [ PASS ] :: Command 'umount /dev/md1' (Expected 0, got 0) :: [ 16:04:14 ] :: [ PASS ] :: Command 'mdadm --create --run /dev/md1 --level 10 --metadata 1.2 --raid-devices 5 /dev/loop0 /dev/loop3 /dev/loop1 /dev/loop2 /dev/loop4 --spare-devices 1 /dev/loop5 --chunk 512 --bitmap=internal --bitmap-chunk=64M' (Expected 0, got 0) :: [ 16:04:15 ] :: [ PASS ] :: Command 'cat /proc/mdstat' (Expected 0, got 0) :: [ 16:04:15 ] :: [ PASS ] :: Command 'lsblk' (Expected 0, got 0) :: [ 16:04:16 ] :: [ PASS ] :: Command 'mdadm --detail /dev/md1' (Expected 0, got 0) :: [ 16:04:16 ] :: [ LOG ] :: INFO: Successfully created md raid /dev/md1 :: [ 16:04:27 ] :: [ LOG ] :: mkfs -t xfs /dev/md1 :: [ 16:04:29 ] :: [ PASS ] :: Command 'mount -t xfs /dev/md1 /mnt/md_test ' (Expected 0, got 0) :: [ 16:04:30 ] :: [ PASS ] :: Command 'fstrim -v /mnt/md_test' (Expected 0, got 0) :: [ 16:04:31 ] :: [ PASS ] :: Command 'umount /dev/md1' (Expected 0, got 0) :: [ 16:04:53 ] :: [ PASS ] :: check the errors (Expected 1, got 1) :::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::: :: Duration: 605s :: Assertions: 47 good, 0 bad :: RESULT: PASS (Test) :::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::: :: unknown :::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::: :: [ 16:04:59 ] :: [ LOG ] :: Phases fingerprint: L5rLAvqh :: [ 16:04:59 ] :: [ LOG ] :: Asserts fingerprint: IcvkrVvU Uploading journal.xml .done :::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::::: :: Duration: 613s :: Phases: 1 good, 0 bad :: OVERALL RESULT: PASS (unknown)