path: glusterfs_11.2-3/tests/utils/pidof.py return False def pidof(processname): for p in psutil.process_iter(): if p.pid == 0: path: glusterfs_11.2-3/tests/utils/pidof.py return 1 try: pidof(argv[1]) except Exception as err: print(err) path: glusterfs_11.2-3/tests/basic/ios-dump.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 3 $H0:$B0/${V0}{0,1,2} TEST $CLI volume set $V0 diagnostics.stats-dump-interval 5 path: glusterfs_11.2-3/tests/basic/ec/ec-data-heal.t cleanup TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 disperse 3 redundancy 1 $H0:$B0/${V0}{0..2} TEST $CLI volume start $V0 path: glusterfs_11.2-3/tests/basic/afr/self-heal.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/brick{0,1} TEST $CLI volume set $V0 stat-prefetch off path: glusterfs_11.2-3/tests/basic/afr/split-brain-healing-ctime.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/${V0}{1,2,3,4} TEST $CLI volume set $V0 cluster.self-heal-daemon off path: glusterfs_11.2-3/tests/bugs/replicate/bug-1363721.t } TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 3 $H0:$B0/${V0}{0,1,2} TEST $CLI volume set $V0 performance.write-behind off path: glusterfs_11.2-3/tests/bugs/replicate/bug-1722507-type-mismatch-error-handling.t ## Start and create a volume TEST glusterd; TEST pidof glusterd; TEST $CLI volume info; path: glusterfs_11.2-3/tests/bugs/replicate/do-not-reopen-fd.t TEST glusterd; TEST pidof glusterd TEST $CLI volume create $V0 replica 3 $H0:$B0/${V0}{0,1,2} path: glusterfs_11.2-3/tests/basic/distribute/sparse_file_rebalance.t # Start glusterd TEST glusterd; TEST pidof glusterd; TEST $CLI volume info; path: glusterfs_11.2-3/tests/features/ssl-authz.t TEST glusterd TEST pidof glusterd TEST $CLI volume info; path: glusterfs_11.2-3/tests/basic/fuse/active-io-graph-switch.t cleanup; TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 3 $H0:$B0/${V0}{0,1,2} TEST $CLI volume set $V0 flush-behind off path: glusterfs_11.2-3/tests/bugs/io-cache/bug-858242.t TEST glusterd TEST pidof glusterd TEST $CLI volume info; path: glusterfs_11.2-3/tests/basic/acl/acl-inherit.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 $H0:$B0/${V0}{0,1,2} TEST $CLI volume start $V0 path: glusterfs_11.2-3/tests/line-coverage/quiesce-coverage.t TEST glusterd TEST pidof glusterd TEST $CLI volume info; path: glusterfs_11.2-3/tests/bugs/snapshot/bug-1113975-zfs.t TEST glusterd; TEST pidof glusterd; TEST $CLI volume create $V0 $H0:$L1 $H0:$L2 $H0:$L3; path: glusterfs_11.2-3/tests/basic/namespace.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 3 $H0:$B0/${V0}{0,1,2,3,4,5,6,7,8} TEST $CLI volume set $V0 nfs.disable off path: glusterfs_11.2-3/tests/bugs/distribute/bug-882278.t TEST glusterd TEST pidof glusterd TEST $CLI volume info; path: glusterfs_11.2-3/tests/basic/afr/read-subvol-data.t #Init TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/brick{0,1} TEST $CLI volume set $V0 self-heal-daemon off path: glusterfs_11.2-3/tests/bugs/core/bug-1135514-allow-setxattr-with-null-value.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 $H0:$B0/${V0}0 TEST $CLI volume start $V0 path: glusterfs_11.2-3/tests/bugs/fuse/many-groups-for-acl.t TEST glusterd TEST pidof glusterd TEST $CLI volume create ${V0} ${H0}:${B0}/${V0}1 TEST $CLI volume set $V0 nfs.disable off path: glusterfs_11.2-3/tests/bugs/distribute/bug-862967.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 $H0:$B0/${V0}0 $H0:$B0/${V0}1 $H0:$B0/${V0}2 path: glusterfs_11.2-3/tests/bugs/fuse/bug-983477.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 $H0:$B0/${V0} TEST $CLI volume start $V0 path: glusterfs_11.2-3/tests/bugs/shard/bug-1738419.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 3 $H0:$B0/${V0}{0,1,2} TEST $CLI volume set $V0 features.shard on path: glusterfs_11.2-3/tests/bugs/distribute/bug-860663.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 $H0:$B0/${V0}0 $H0:$B0/${V0}1 $H0:$B0/${V0}2 path: glusterfs_11.2-3/tests/bugs/distribute/bug-1086228.t ## Start and create a volume TEST glusterd; TEST pidof glusterd; TEST $CLI volume create $V0 $H0:$B0/${V0}{1,2} TEST $CLI volume start $V0; path: glusterfs_11.2-3/tests/volume.rc then local pid=$(cat ${rundir}/${V0}.pid) pidof glusterfs 2>&1 | grep -w $pid > /dev/null if [ $? -eq 0 ] path: glusterfs_11.2-3/tests/bugs/replicate/bug-1761531-metadata-heal-restore-time.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 3 $H0:$B0/brick{0..2} TEST $CLI volume start $V0 path: glusterfs_11.2-3/tests/bugs/core/bug-834465.t TEST glusterd TEST pidof glusterd TEST $CLI volume info; path: glusterfs_11.2-3/tests/basic/distribute/directory-rebalance.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 $H0:$B0/${V0}0 TEST $CLI volume set $V0 performance.quick-read off path: glusterfs_11.2-3/tests/basic/ec/ec-up.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 disperse-data 2 redundancy 1 $H0:$B0/${V0}{0,1,3,4,5,6} TEST $CLI volume start $V0 path: glusterfs_11.2-3/tests/bugs/core/bug-1117951.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 $H0:$B0/brick EXPECT 'Created' volinfo_field $V0 'Status'; path: glusterfs_11.2-3/tests/bugs/core/brick-mux-fd-cleanup.t cleanup; TEST glusterd TEST pidof glusterd function keep_fd_open { path: glusterfs_11.2-3/tests/basic/afr/arbiter-add-brick.t TEST glusterd TEST pidof glusterd #Create replica 2 volume and create file/dir. path: glusterfs_11.2-3/extras/init.d/glustereventsd-Debian.in do_status() { pidofproc -p $PIDFILE $DAEMON >/dev/null status=$? if [ $status -eq 0 ]; then path: glusterfs_11.2-3/extras/init.d/glustereventsd-Debian.in do_start() { pidofproc -p $PIDFILE $DAEMON >/dev/null status=$? if [ $status -eq 0 ]; then path: glusterfs_11.2-3/tests/gfid2path/get-gfid-to-path.t TEST glusterd TEST pidof glusterd ## Create a 1*2 volume path: glusterfs_11.2-3/tests/bugs/bitrot/bug-1228680.t ## Start glusterd TEST glusterd; TEST pidof glusterd; ## Lets create and start the volume path: glusterfs_11.2-3/tests/bugs/glusterd/removing-multiple-bricks-in-single-remove-brick-command.t #Basic checks TEST glusterd TEST pidof glusterd TEST $CLI volume info path: glusterfs_11.2-3/tests/basic/ctime/ctime-heal-symlinks.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 disperse 3 redundancy 1 $H0:$B0/${V0}{0,1,2} TEST $CLI volume set $V0 performance.stat-prefetch off path: glusterfs_11.2-3/tests/basic/ctime/ctime-heal-symlinks.t ############################################################################### TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 3 $H0:$B0/${V0}{0,1,2} TEST $CLI volume set $V0 performance.stat-prefetch off path: glusterfs_11.2-3/tests/bugs/snapshot/bug-1109889-zfs.t TEST glusterd; TEST pidof glusterd; TEST $CLI volume create $V0 $H0:$L1 $H0:$L2 $H0:$L3; path: glusterfs_11.2-3/tests/bugs/snapshot/bug-1109770-zfs.t TEST glusterd; TEST pidof glusterd; TEST $CLI volume create $V0 $H0:$L1 $H0:$L2 $H0:$L3; path: glusterfs_11.2-3/tests/features/simple-quota-dht.t TEST glusterd TEST pidof glusterd TEST $CLI volume info; path: glusterfs_11.2-3/tests/bugs/gfapi/bug-1319374-THIS-crash.t TEST glusterd TEST pidof glusterd TEST $CLI volume info; path: glusterfs_11.2-3/tests/basic/ec/ec-optimistic-changelog.t cleanup TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 disperse 3 redundancy 1 $H0:$B0/${V0}{0..2} TEST $CLI volume heal $V0 disable path: glusterfs_11.2-3/tests/basic/inode-namespace.t TEST glusterd TEST pidof glusterd TEST $CLI volume info; path: glusterfs_11.2-3/tests/basic/quota.t TEST glusterd TEST pidof glusterd TEST $CLI volume info; path: glusterfs_11.2-3/tests/basic/ctime/ctime-ec-heal.t #cleate and start volume TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 disperse 3 redundancy 1 $H0:$B0/${V0}{1..3} TEST $CLI volume start $V0 path: glusterfs_11.2-3/tests/basic/glusterd/thin-arbiter-volume.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 thin-arbiter 1 $H0:$B0/b1 $H0:$B0/b2 $H0:$B0/b3 path: glusterfs_11.2-3/tests/basic/glusterd/thin-arbiter-volume.t TEST killall -15 glusterd TEST glusterd TEST pidof glusterd EXPECT "1 x 2 = 2" volinfo_field $V0 "Number of Bricks" path: glusterfs_11.2-3/tests/bugs/snapshot/bug-1064768.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/brick0 $H0:$B0/brick1 TEST $CLI volume start $V0 path: glusterfs_11.2-3/tests/bugs/cli/bug-1378842-volume-get-all.t cleanup; TEST glusterd TEST pidof glusterd TEST $CLI volume set all server-quorum-ratio 80 path: glusterfs_11.2-3/tests/features/worm.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 $H0:$B0/${V0}1 path: glusterfs_11.2-3/tests/basic/volume-status.t TEST glusterd TEST pidof glusterd TEST $CLI volume info; path: glusterfs_11.2-3/tests/basic/mount-options.disabled TEST glusterd --xlator-option=*.rpc-auth-allow-insecure=on TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/${V0}{0,1,2,3,4,5} TEST $CLI volume set $V0 server.allow-insecure on path: glusterfs_11.2-3/tests/features/ipc.t TEST glusterd TEST pidof glusterd TEST $CLI volume info; path: glusterfs_11.2-3/tests/line-coverage/cli-peer-and-volume-operations.t # try unrecognised command TEST ! $CLI_1 volume TEST pidof glusterd ## all help commands path: glusterfs_11.2-3/tests/bugs/cli/bug-921215.t TEST glusterd TEST pidof glusterd TEST ! $CLI volume create $V0.temp replica 2 $H0:$B0/${V0}0 $H0:$B0/${V0}1 path: glusterfs_11.2-3/tests/bugs/replicate/bug-1130892.t TEST glusterd TEST pidof glusterd TEST $CLI volume info; path: glusterfs_11.2-3/tests/bugs/core/bug-1110917.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 $H0:$B0/brick1 $H0:$B0/brick2; path: glusterfs_11.2-3/tests/bugs/glusterfs-server/bug-904300.t # 1-8 TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 $H0:$B0/$V0; path: glusterfs_11.2-3/tests/basic/ctime/ctime-rep-rebalance.t #cleate and start volume TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 3 $H0:$B0/${V0}{0..5} TEST $CLI volume start $V0 path: glusterfs_11.2-3/tests/basic/open-behind/open-behind.t TEST glusterd TEST pidof glusterd TEST ${CLI} volume create ${V0} ${H0}:${B0}/${V0} TEST ${CLI} volume set ${V0} flush-behind off path: glusterfs_11.2-3/tests/basic/afr/rename-data-loss.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/brick{0,1} TEST $CLI volume set $V0 write-behind off path: glusterfs_11.2-3/tests/bugs/core/bug-903336.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 $H0:$B0/${V0} TEST $CLI volume start $V0 path: glusterfs_11.2-3/tests/bugs/glusterfs/bug-893338.t ## Start and create a volume TEST glusterd; TEST pidof glusterd; TEST $CLI volume info; path: glusterfs_11.2-3/tests/bugs/replicate/bug-1480525.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/${V0}{0,1} TEST $CLI volume start $V0 path: glusterfs_11.2-3/tests/bugs/core/bug-927616.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/${V0}{1,2}; TEST $CLI volume set $V0 performance.open-behind off; path: glusterfs_11.2-3/tests/bugs/bug-1371806.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 $H0:$B0/${V0}{0,1,2,3,4,5} TEST $CLI volume start $V0 path: glusterfs_11.2-3/tests/bugs/glusterd/daemon-log-level-option.t # Basic checks TEST glusterd TEST pidof glusterd TEST $CLI volume info path: glusterfs_11.2-3/tests/bugs/posix/bug-1034716.t #Basic checks TEST glusterd TEST pidof glusterd TEST $CLI volume info path: glusterfs_11.2-3/tests/bugs/protocol/bug-808400-repl.t TEST glusterd TEST pidof glusterd TEST $CLI volume info; TEST $CLI volume create $V0 replica 2 $H0:$B0/brick1 $H0:$B0/brick2; path: glusterfs_11.2-3/tests/bugs/replicate/bug-1637249-gfid-heal.t TEST glusterd; TEST pidof glusterd; TEST $CLI volume create $V0 replica 2 $H0:$B0/${V0}{0,1}; TEST $CLI volume set $V0 self-heal-daemon off path: glusterfs_11.2-3/tests/basic/ec/ec-stripe.t cleanup TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 disperse 3 redundancy 1 $H0:$B0/${V0}{0..2} TEST $CLI volume heal $V0 disable path: glusterfs_11.2-3/tests/bugs/replicate/bug-976800.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/${V0}{0,1} TEST $CLI volume set $V0 cluster.ensure-durability off path: glusterfs_11.2-3/tests/basic/afr/halo.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/${V0}{0,1} TEST $CLI volume set $V0 cluster.halo-enabled yes path: glusterfs_11.2-3/tests/basic/ec/ec-background-heals.t cleanup TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 disperse 3 redundancy 1 $H0:$B0/${V0}{0..2} TEST $CLI volume heal $V0 disable path: glusterfs_11.2-3/tests/bugs/replicate/bug-1717819-metadata-split-brain-detection.t ## Start and create a volume TEST glusterd; TEST pidof glusterd; TEST $CLI volume info; path: glusterfs_11.2-3/tests/bugs/snapshot/bug-1275616.t TEST verify_lvm_version; TEST glusterd; TEST pidof glusterd; TEST setup_lvm 1 path: glusterfs_11.2-3/tests/basic/afr/heal-quota.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/${V0}{0,1} TEST $CLI volume set $V0 cluster.self-heal-daemon off path: glusterfs_11.2-3/tests/bugs/shard/bug-1245547.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 $H0:$B0/${V0}{0,1} TEST $CLI volume set $V0 features.shard on path: glusterfs_11.2-3/tests/basic/nufa.t TEST glusterd TEST pidof glusterd TEST $CLI volume info; path: glusterfs_11.2-3/tests/basic/afr/granular-esh/conservative-merge.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/${V0}{0,1} TEST $CLI volume start $V0 path: glusterfs_11.2-3/tests/bugs/glusterfs/bug-811493.t TEST glusterd TEST pidof glusterd TEST $CLI system uuid reset; path: glusterfs_11.2-3/tests/bugs/bitrot/bug-1288490.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/brick{0,1} path: glusterfs_11.2-3/tests/bugs/snapshot/bug-1168875.t TEST glusterd; TEST pidof glusterd; TEST $CLI volume create $V0 $H0:$L1 $H0:$L2 $H0:$L3; path: glusterfs_11.2-3/tests/bugs/cli/bug-1320388.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 disperse 6 redundancy 2 $H0:$B0/${V0}{0..5} TEST $CLI volume set $V0 disperse.eager-lock off path: glusterfs_11.2-3/tests/bugs/snapshot/bug-1322772-real-path-fix-for-snapshot-zfs.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 $H0:$L1 path: glusterfs_11.2-3/tests/bugs/snapshot/bug-1322772-real-path-fix-for-snapshot-zfs.t TEST glusterd TEST pidof glusterd EXPECT_WITHIN $PROCESS_UP_TIMEOUT "1" brick_up_status $V0 $H0 $L1 path: glusterfs_11.2-3/extras/init.d/glusterd-Debian.in do_start() { pidofproc -p $PIDFILE $DAEMON >/dev/null status=$? if [ $status -eq 0 ]; then path: glusterfs_11.2-3/extras/init.d/glusterd-Debian.in do_status() { pidofproc -p $PIDFILE $DAEMON >/dev/null status=$? if [ $status -eq 0 ]; then path: glusterfs_11.2-3/tests/basic/posix/zero-fill-enospace.t TEST glusterd; TEST pidof glusterd; TEST truncate -s 300M $B0/brick1 path: glusterfs_11.2-3/tests/basic/ec/ec-dirty-flags.t cleanup TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 disperse 3 redundancy 1 $H0:$B0/${V0}{0..2} TEST $CLI volume heal $V0 disable path: glusterfs_11.2-3/tests/bugs/snapshot/bug-1157991.t TEST verify_lvm_version; TEST glusterd; TEST pidof glusterd; TEST setup_lvm 1 path: glusterfs_11.2-3/tests/000-flaky/bugs_glusterd_quorum-value-check.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 3 $H0:$B0/${V0}{0,1,2} TEST $CLI volume set $V0 nfs.disable off path: glusterfs_11.2-3/tests/00-geo-rep/georep-basic-dr-rsync-arbiter.t cleanup; TEST glusterd; TEST pidof glusterd path: glusterfs_11.2-3/tests/bugs/fuse/bug-1283103.t #Basic checks TEST glusterd TEST pidof glusterd TEST $CLI volume info path: glusterfs_11.2-3/tests/basic/volume-snap-scheduler.t TEST glusterd; TEST pidof glusterd; TEST $CLI volume create $V0 replica 2 $H0:$B0/${GMV0}{1,2,3,4}; path: glusterfs_11.2-3/tests/basic/ec/ec-notify.t cleanup TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 disperse 3 redundancy 1 $H0:$B0/${V0}{0..2} TEST $CLI volume start $V0 path: glusterfs_11.2-3/tests/00-geo-rep/georep-basic-tarssh-ec.t cleanup; TEST glusterd; TEST pidof glusterd path: glusterfs_11.2-3/tests/bugs/core/bug-1168803-snapd-option-validation-fix.t ## Start glusterd TEST glusterd; TEST pidof glusterd; ## create volume path: glusterfs_11.2-3/tests/basic/afr/arbiter-mount.t TEST glusterd; TEST pidof glusterd TEST $CLI volume create $V0 replica 3 arbiter 1 $H0:$B0/${V0}{0,1,2} path: glusterfs_11.2-3/tests/bugs/gfapi/bug-1093594.t ## Start and create a volume TEST glusterd; TEST pidof glusterd; TEST $CLI volume info; path: glusterfs_11.2-3/tests/basic/ec/nfs.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 redundancy 2 $H0:$B0/${V0}{0..5} EXPECT "Created" volinfo_field $V0 'Status' path: glusterfs_11.2-3/tests/basic/afr/afr-up.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 3 $H0:$B0/${V0}{0,1,3,4,5,6} TEST $CLI volume start $V0 path: glusterfs_11.2-3/tests/bugs/replicate/bug-1655052-sbrain-policy-same-size.t TEST glusterd TEST pidof glusterd #Create replica 2 volume path: glusterfs_11.2-3/extras/test/bug-920583.t ## Start and create a volume TEST glusterd; TEST pidof glusterd; TEST $CLI volume create $V0 replica 2 stripe 2 $H0:$B0/${V0}{1,2,3,4,5,6,7,8}; path: glusterfs_11.2-3/tests/bugs/replicate/mdata-heal-no-xattrs.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 3 $H0:$B0/${V0}{0,1,2}; TEST $CLI volume set $V0 cluster.self-heal-daemon off path: glusterfs_11.2-3/tests/bugs/shard/issue-2038.t TEST verify_lvm_version TEST glusterd TEST pidof glusterd TEST init_n_bricks 1 TEST setup_lvm 1 path: glusterfs_11.2-3/tests/bugs/replicate/bug-1305031-block-reads-on-metadata-sbrain.t #Test that for files in metadata-split-brain, we do not wind even a single read. TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/${V0}{0,1} path: glusterfs_11.2-3/tests/line-coverage/meta-max-coverage.t TEST glusterd TEST pidof glusterd TEST $CLI volume info; path: glusterfs_11.2-3/tests/basic/ctime/ctime-ec-rebalance.t #cleate and start volume TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 disperse 3 redundancy 1 $H0:$B0/${V0}{0..5} TEST $CLI volume start $V0 path: glusterfs_11.2-3/tests/bugs/replicate/bug-1734370-entry-heal-restore-time.t TEST glusterd; TEST pidof glusterd; TEST $CLI volume create $V0 replica 3 $H0:$B0/${V0}{0,1,2}; TEST $CLI volume start $V0; path: glusterfs_11.2-3/tests/bugs/glusterd/bug-949930.t TEST glusterd; TEST pidof glusterd; TEST $CLI volume create $V0 $H0:$B0/${V0}{1,2}; path: glusterfs_11.2-3/tests/bugs/cli/bug-1022905.t ## Create a volume TEST glusterd; TEST pidof glusterd TEST $CLI volume info; path: glusterfs_11.2-3/tests/bugs/shard/bug-shard-zerofill.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/${V0}{0..3} TEST $CLI volume set $V0 features.shard on path: glusterfs_11.2-3/tests/bugs/nfs/zero-atime.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 $H0:$B0/$V0 path: glusterfs_11.2-3/tests/basic/hardlink-limit.t TEST glusterd TEST pidof glusterd TEST $CLI volume info; path: glusterfs_11.2-3/tests/basic/afr/data-self-heal.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/brick{0,1} TEST $CLI volume set $V0 cluster.background-self-heal-count 0 path: glusterfs_11.2-3/tests/bugs/quota/bug-1250582-volume-reset-should-not-remove-quota-quota-deem-statfs.t TEST glusterd; TEST pidof glusterd; TEST $CLI volume info; path: glusterfs_11.2-3/tests/gfid2path/block-mount-access.t TEST glusterd TEST pidof glusterd ## Create a 2*2 volume path: glusterfs_11.2-3/tests/bugs/replicate/bug-859581.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/${V0}{1,2} EXPECT 'Created' volinfo_field $V0 'Status'; path: glusterfs_11.2-3/tests/basic/cloudsync-sanity.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 3 $H0:$B0/${V0}{1,2,3,4,5,6,7,8,9}; path: glusterfs_11.2-3/tests/bugs/snapshot/bug-1597662.t TEST setup_lvm 3; TEST glusterd; TEST pidof glusterd; TEST $CLI volume create $V0 $H0:$L1 $H0:$L2 $H0:$L3; path: glusterfs_11.2-3/extras/init.d/glustereventsd-Redhat.in { echo -n $"Stopping $BASE:" if pidofproc -p $PIDFILE $GLUSTEREVENTSD_BIN &> /dev/null; then killproc -p $PIDFILE $BASE else path: glusterfs_11.2-3/extras/init.d/glustereventsd-Redhat.in start() { if pidofproc -p $PIDFILE $GLUSTEREVENTSD_BIN &> /dev/null; then echo "glustereventsd service is already running with pid $PID" return 0 path: glusterfs_11.2-3/tests/bugs/write-behind/issue-884.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 $H0:$B0/$V0 path: glusterfs_11.2-3/tests/bitrot/bug-1221914.t TEST glusterd TEST pidof glusterd ## Create a distribute volume (B=2) path: glusterfs_11.2-3/tests/bugs/snapshot/bug-1109889.t TEST glusterd; TEST pidof glusterd; TEST $CLI volume create $V0 $H0:$L1 $H0:$L2 $H0:$L3; path: glusterfs_11.2-3/tests/bugs/distribute/bug-1042725.t #Basic checks TEST glusterd TEST pidof glusterd TEST $CLI volume info path: glusterfs_11.2-3/tests/bugs/geo-replication/bug-1111490.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 $H0:$B0/${V0}0 path: glusterfs_11.2-3/tests/bugs/glusterd/bug-824753.t ## Start and create a volume TEST glusterd; TEST pidof glusterd; TEST $CLI volume info; path: glusterfs_11.2-3/tests/00-geo-rep/georep-basic-dr-tarssh-arbiter.t cleanup; TEST glusterd; TEST pidof glusterd path: glusterfs_11.2-3/tests/basic/fops-sanity.t TEST glusterd TEST pidof glusterd TEST $CLI volume info; path: glusterfs_11.2-3/tests/bugs/fuse/setup.sh #Basic checks TEST glusterd TEST pidof glusterd TEST $CLI volume info path: glusterfs_11.2-3/tests/basic/gfid-access.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 $H0:$B0/${V0}0 TEST $CLI volume start $V0 path: glusterfs_11.2-3/tests/bugs/shard/issue-1384.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 3 $H0:$B0/${V0}{0,1,3} TEST $CLI volume set $V0 features.shard on path: glusterfs_11.2-3/tests/basic/ec/ec-fix-openfd.t cleanup TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 disperse 3 redundancy 1 $H0:$B0/${V0}{0..2} TEST $CLI volume set $V0 performance.read-after-open yes path: glusterfs_11.2-3/tests/bugs/replicate/bug-913051.t # open-fd-fix in afr. TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/${V0}0 $H0:$B0/${V0}1 TEST $CLI volume set $V0 cluster.self-heal-daemon off path: glusterfs_11.2-3/tests/bugs/shard/bug-1272986.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/${V0}{0,1,2,3} TEST $CLI volume set $V0 features.shard on path: glusterfs_11.2-3/tests/bugs/write-behind/bug-1279730.t ## Start and create a volume TEST glusterd TEST pidof glusterd; TEST $CLI volume info; path: glusterfs_11.2-3/tests/basic/afr/name-self-heal.t #Check that when quorum is enabled name-heal happens as expected TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 3 $H0:$B0/brick{0,1,2} TEST $CLI volume set $V0 performance.stat-prefetch off path: glusterfs_11.2-3/tests/basic/afr/name-self-heal.t #Check that when quorum is not enabled name-heal happens correctly TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/brick{0,1} TEST $CLI volume set $V0 performance.stat-prefetch off path: glusterfs_11.2-3/tests/basic/afr/split-brain-open.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/${V0}{0,1} TEST $CLI volume start $V0 path: glusterfs_11.2-3/tests/bugs/fuse/bug-1336818.t # Prepare TEST glusterd TEST pidof glusterd TEST $CLI volume info; path: glusterfs_11.2-3/tests/bugs/cli/bug-1047378.t TEST glusterd TEST pidof glusterd TEST "echo volume list | $CLI --xml | xmllint --format -" path: glusterfs_11.2-3/tests/basic/afr/afr-anon-inode-no-quorum.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/${V0}{0,1} TEST $CLI volume heal $V0 disable path: glusterfs_11.2-3/tests/bugs/nfs/bug-1210338.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 $H0:$B0/$V0 path: glusterfs_11.2-3/tests/basic/ctime/ctime-utimesat.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 3 $H0:$B0/${V0}{0,1,2} TEST $CLI volume set $V0 performance.stat-prefetch off path: glusterfs_11.2-3/tests/bugs/glusterfs/bug-873962-spb.t cleanup; TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/${V0}{0,1} path: glusterfs_11.2-3/tests/bugs/shard/bug-1605056-2.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 3 $H0:$B0/${V0}{0,1,2} TEST $CLI volume set $V0 features.shard on path: glusterfs_11.2-3/tests/bugs/distribute/bug-907072.t TEST glusterd; TEST pidof glusterd; TEST $CLI volume create $V0 $H0:$B0/${V0}{0,1,2,3}; path: glusterfs_11.2-3/tests/bugs/shard/bug-1261773.t TEST glusterd TEST pidof glusterd TEST check_option_help_presence "features.shard" path: glusterfs_11.2-3/tests/bugs/replicate/bug-921231.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/${V0}0 $H0:$B0/${V0}1 TEST $CLI volume set $V0 cluster.eager-lock on path: glusterfs_11.2-3/tests/bugs/shard/issue-1243.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 $H0:$B0/${V0}{0,1} TEST $CLI volume set $V0 features.shard on path: glusterfs_11.2-3/tests/bugs/trace/bug-797171.t TEST glusterd TEST pidof glusterd TEST $CLI volume info; path: glusterfs_11.2-3/tests/bugs/nfs/bug-974972.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/${V0}{0,1} TEST $CLI volume set $V0 self-heal-daemon off path: glusterfs_11.2-3/tests/bugs/shard/bug-1468483.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 $H0:$B0/${V0}0 TEST $CLI volume set $V0 performance.write-behind off path: glusterfs_11.2-3/tests/bugs/replicate/bug-1266876-allow-reset-brick-for-same-path.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/${V0}{0,1} TEST $CLI volume start $V0 path: glusterfs_11.2-3/tests/bugs/core/bug-845213.t ## Start and create a volume TEST glusterd; TEST pidof glusterd; TEST $CLI volume info; path: glusterfs_11.2-3/tests/bugs/cli/bug-983317-volume-get.t cleanup; TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 $H0:$B0/${V0}{1,2}; path: glusterfs_11.2-3/tests/bugs/glusterd/859927/repl.t TEST glusterd; TEST pidof glusterd #Tests for data-self-heal-algorithm option path: glusterfs_11.2-3/tests/00-geo-rep/georep-basic-dr-tarssh.t cleanup; TEST glusterd; TEST pidof glusterd path: glusterfs_11.2-3/tests/00-geo-rep/bug-1600145.t SCRIPT_TIMEOUT=600 TEST glusterd; TEST pidof glusterd ##Variables path: glusterfs_11.2-3/tests/bugs/bitrot/1209818-vol-info-show-scrub-process-properly.t ## Start glusterd TEST glusterd; TEST pidof glusterd; ## Lets create and start the volume path: glusterfs_11.2-3/tests/bugs/bug-1371806_acl.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 $H0:$B0/${V0}{0,1,2,3,4,5} TEST $CLI volume set $V0 diagnostics.client-log-level DEBUG path: glusterfs_11.2-3/tests/bugs/glusterfs-server/bug-861542.t TEST glusterd TEST pidof glusterd TEST $CLI volume info; # Distributed volume with a single brick was chosen solely for the ease of path: glusterfs_11.2-3/tests/bugs/replicate/bug-1325792.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/${V0}{0,1,2,3} TEST $CLI volume start $V0 path: glusterfs_11.2-3/tests/basic/afr/read-subvol-entry.t #Init TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/brick{0,1} TEST $CLI volume set $V0 self-heal-daemon off path: glusterfs_11.2-3/tests/features/delay-gen.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 $H0:$B0/${V0}1 path: glusterfs_11.2-3/tests/basic/afr/gfid-mismatch-resolution-with-fav-child-policy.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/${V0}{0,1} TEST $CLI volume start $V0 path: glusterfs_11.2-3/tests/bugs/ec/bug-1236065.t TEST ec_test_make TEST pidof glusterd EXPECT "$V0" volinfo_field $V0 'Volume Name' EXPECT 'Started' volinfo_field $V0 'Status' path: glusterfs_11.2-3/tests/bugs/ec/bug-1236065.t ## step 1 TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 disperse 7 redundancy 3 $H0:$B0/${V0}{0..6} TEST $CLI volume start $V0 path: glusterfs_11.2-3/tests/basic/ec/ec-1468261.t #cleate and start volume TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 disperse 6 redundancy 2 $H0:$B0/${V0}{0..5} TEST $CLI volume set $V0 disperse.optimistic-change-log on path: glusterfs_11.2-3/tests/line-coverage/cli-negative-case-and-function-coverage.t # Start glusterd TEST glusterd; TEST pidof glusterd; # Checking various negative cases of gluster cli path: glusterfs_11.2-3/tests/00-geo-rep/georep-basic-rsync-ec.t cleanup; TEST glusterd; TEST pidof glusterd path: glusterfs_11.2-3/tests/basic/ec/statedump.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 redundancy 1 $H0:$B0/${V0}{0..2} EXPECT "Created" volinfo_field $V0 'Status' path: glusterfs_11.2-3/tests/basic/ec/quota.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 disperse $H0:$B0/${V0}{0..2} EXPECT 'Created' volinfo_field $V0 'Status' path: glusterfs_11.2-3/tests/basic/ec/ec-discard.t #cleate and start volume TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 disperse 6 redundancy 2 $H0:$B0/${V0}{0..5} TEST $CLI volume set $V0 disperse.optimistic-change-log on path: glusterfs_11.2-3/tests/bugs/replicate/bug-1801624-entry-heal.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 3 $H0:$B0/brick{0,1,2} TEST $CLI volume set $V0 heal-timeout 5 path: glusterfs_11.2-3/tests/bugs/fuse/bug-858488-min-free-disk.t ## Start glusterd TEST glusterd; TEST pidof glusterd; TEST $CLI volume info; path: glusterfs_11.2-3/tests/bugs/replicate/bug-1297695.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/${V0}0 $H0:$B0/${V0}1 path: glusterfs_11.2-3/tests/bugs/ec/bug-1699866-check-reopen-fd.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 disperse 6 redundancy 2 $H0:$B0/${V0}{0..5} TEST $CLI volume heal $V0 disable path: glusterfs_11.2-3/tests/bugs/glusterfs/bug-896431.t ## Start and create a volume TEST glusterd; TEST pidof glusterd; TEST $CLI volume create $V0 replica 3 $H0:$B0/${V0}{1,2,3,4,5,6}; path: glusterfs_11.2-3/tests/bugs/glusterfs/bug-896431.t ## Start and create a pure replicate volume TEST glusterd; TEST pidof glusterd; TEST $CLI volume create $V0 replica 8 $H0:$B0/${V0}{1,2,3,4,5,6,7,8}; path: glusterfs_11.2-3/tests/basic/quota-nfs.t TEST glusterd TEST pidof glusterd TEST $CLI volume info; path: glusterfs_11.2-3/tests/bugs/bitrot/bug-1227996.t ## Start glusterd TEST glusterd; TEST pidof glusterd; ## Lets create and start the volume path: glusterfs_11.2-3/tests/bugs/ec/bug-1187474.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 redundancy 2 $H0:$B0/${V0}{0..5} EXPECT "Created" volinfo_field $V0 'Status' path: glusterfs_11.2-3/tests/bugs/replicate/bug-802417.t TEST glusterd TEST pidof glusterd TEST $CLI volume info; path: glusterfs_11.2-3/tests/basic/glusterd/heald.t cleanup; TEST glusterd TEST pidof glusterd #Commands should fail when volume doesn't exist path: glusterfs_11.2-3/tests/bugs/ec/bug-1179050.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 redundancy 1 $H0:$B0/${V0}{0..2} EXPECT "Created" volinfo_field $V0 'Status' path: glusterfs_11.2-3/tests/basic/afr/split-brain-resolution.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/${V0}{0,1} TEST $CLI volume set $V0 cluster.data-self-heal on path: glusterfs_11.2-3/tests/bugs/glusterd/bug-1070734.t ## Start glusterd TEST glusterd; TEST pidof glusterd; TEST $CLI volume info; path: glusterfs_11.2-3/tests/basic/ec/lock-contention.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 disperse 3 redundancy 1 $H0:$B0/${V0}{0..2} TEST $CLI volume set $V0 performance.stat-prefetch off path: glusterfs_11.2-3/tests/bugs/posix/bug-1040275-brick-uid-reset-on-volume-restart.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 3 $H0:$B0/${V0}{1,2,3,4,5,6}; path: glusterfs_11.2-3/tests/bugs/replicate/bug-1438255-do-not-mark-self-accusing-xattrs.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 3 $H0:$B0/${V0}{0,1,2} TEST $CLI volume set $V0 cluster.self-heal-daemon off path: glusterfs_11.2-3/tests/bugs/cli/bug-822830.t ## Start and create a volume TEST glusterd; TEST pidof glusterd; TEST $CLI volume create $V0 replica 3 $H0:$B0/${V0}{1,2,3,4,5,6}; path: glusterfs_11.2-3/tests/bugs/shard/bug-1488546.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 3 $H0:$B0/${V0}{0,1,2} TEST $CLI volume set $V0 features.shard on path: glusterfs_11.2-3/tests/bugs/heal-symlinks.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 disperse 3 redundancy 1 $H0:$B0/${V0}{0,1,2} TEST $CLI volume set $V0 performance.stat-prefetch off path: glusterfs_11.2-3/tests/bugs/heal-symlinks.t ############################################################################### TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 3 $H0:$B0/${V0}{0,1,2} TEST $CLI volume set $V0 performance.stat-prefetch off path: glusterfs_11.2-3/tests/bugs/geo-replication/bug-877293.t TEST glusterd TEST pidof glusterd ## Start and create a replicated volume path: glusterfs_11.2-3/tests/bugs/replicate/issue-3238-mdata-unnecessary-heal.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 arbiter 1 $H0:$B0/${V0}{0,1,2} TEST $CLI volume set $V0 cluster.self-heal-daemon off path: glusterfs_11.2-3/tests/bugs/bug-1584517.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 3 $H0:$B0/${V0}{0,1,2,3,4,5} TEST $CLI volume start $V0 path: glusterfs_11.2-3/tests/line-coverage/arbiter-coverage.t ## Start and create a volume TEST glusterd; TEST pidof glusterd; TEST $CLI volume info; path: glusterfs_11.2-3/tests/basic/volume.t TEST glusterd TEST pidof glusterd TEST $CLI volume info; path: glusterfs_11.2-3/tests/bugs/distribute/bug-1543279.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 $H0:$B0/${V0}{0..8} force path: glusterfs_11.2-3/tests/000-flaky/basic_ec_ec-quorum-count-partial-failure.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 disperse 6 redundancy 2 $H0:$B0/${V0}{0..5} TEST $CLI volume create $V1 $H0:$B0/${V1}{0..5} path: glusterfs_11.2-3/tests/basic/changelog/changelog-rename.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 $H0:$B0/${V0}0 TEST $CLI volume set $V0 changelog.changelog on path: glusterfs_11.2-3/tests/basic/changelog/changelog-rename.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 $H0:$B0/${V0}{0,1} TEST $CLI volume set $V0 changelog.changelog on path: glusterfs_11.2-3/tests/bugs/logging/bug-823081.t TEST glusterd TEST pidof glusterd logdir=`gluster --print-logdir` path: glusterfs_11.2-3/tests/bugs/replicate/bug-1335652.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 3 $H0:$B0/${V0}{0,1,2} TEST $CLI volume set $V0 shard on path: glusterfs_11.2-3/tests/000-flaky/bugs_distribute_bug-1117851.t TEST glusterd TEST pidof glusterd TEST $CLI volume info; path: glusterfs_11.2-3/tests/bugs/replicate/bug-765564.t TEST glusterd TEST pidof glusterd ## Start and create a volume path: glusterfs_11.2-3/tests/basic/ec/ec-fallocate.t #cleate and start volume TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 disperse 3 redundancy 1 $H0:$B0/${V0}{0..2} TEST $CLI volume start $V0 path: glusterfs_11.2-3/tests/bugs/shard/bug-1260637.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 $H0:$B0/${V0}0 TEST $CLI volume set $V0 features.shard on path: glusterfs_11.2-3/tests/basic/afr/gfid-heal.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/brick{0,1,2,3,4,5} TEST $CLI volume start $V0 path: glusterfs_11.2-3/tests/bugs/replicate/bug-1037501.t cleanup; TEST glusterd TEST pidof glusterd TEST $CLI volume info; path: glusterfs_11.2-3/tests/bugs/replicate/bug-1340623-mkdir-fails-remove-brick-started.t TEST glusterd TEST pidof glusterd ## Create a 2x2 volume path: glusterfs_11.2-3/tests/bugs/posix/bug-1619720.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 $H0:$B0/${V0}0 $H0:$B0/${V0}1 path: glusterfs_11.2-3/tests/bugs/ec/bug-1188145.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 redundancy 2 $H0:$B0/${V0}{0..5} EXPECT "Created" volinfo_field $V0 'Status' path: glusterfs_11.2-3/tests/basic/shd-mux-ec.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 3 $H0:$B0/${V0}{0,1,2,3,4,5} TEST $CLI volume set $V0 cluster.background-self-heal-count 0 path: glusterfs_11.2-3/tests/basic/shd-mux-afr.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 3 $H0:$B0/${V0}{0,1,2,3,4,5} TEST $CLI volume set $V0 cluster.background-self-heal-count 0 path: glusterfs_11.2-3/tests/bugs/md-cache/afr-stale-read.t #Basic checks TEST glusterd TEST pidof glusterd TEST $CLI volume info path: glusterfs_11.2-3/tests/00-geo-rep/georep-basic-dr-rsync.t cleanup; TEST glusterd; TEST pidof glusterd path: glusterfs_11.2-3/tests/bugs/snapshot/bug-1232430.t TEST verify_lvm_version; TEST glusterd -LDEBUG; TEST pidof glusterd; TEST setup_lvm 1 path: glusterfs_11.2-3/tests/bugs/replicate/bug-865825.t TEST glusterd TEST pidof glusterd TEST $CLI volume info; path: glusterfs_11.2-3/tests/bugs/replicate/bug-918437-sh-mtime.t ## Tests if mtime is correct after self-heal. TEST glusterd TEST pidof glusterd TEST mkdir -p $B0/gfs0/brick0{1,2} TEST $CLI volume create $V0 replica 2 transport tcp $H0:$B0/gfs0/brick01 $H0:$B0/gfs0/brick02 path: glusterfs_11.2-3/tests/basic/ec/ec-seek.t TEST glusterd TEST pidof glusterd TEST $CLI volume info path: glusterfs_11.2-3/tests/bugs/nfs/bug-915280.t TEST glusterd TEST pidof glusterd function volinfo_field() path: glusterfs_11.2-3/tests/bugs/unclassified/bug-874498.t TEST glusterd TEST pidof glusterd TEST $CLI volume info; TEST $CLI volume create $V0 replica 2 $H0:$B0/brick1 $H0:$B0/brick2; path: glusterfs_11.2-3/tests/bitrot/br-stub.t TEST glusterd TEST pidof glusterd ## Create a distribute volume (B=2) path: glusterfs_11.2-3/tests/basic/quick-read-with-upcall.t #Basic checks TEST glusterd TEST pidof glusterd TEST $CLI volume info path: glusterfs_11.2-3/tests/bugs/distribute/issue-1327.t TEST glusterd TEST pidof glusterd BRICK1=$B0/${V0}-0 path: glusterfs_11.2-3/tests/bugs/replicate/bug-1749322-entry-heal-not-happening.t ## Start and create a volume TEST glusterd; TEST pidof glusterd; TEST $CLI volume info; path: glusterfs_11.2-3/tests/bugs/shard/zero-flag.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/${V0}{0,1,2,3} TEST $CLI volume set $V0 features.shard on path: glusterfs_11.2-3/tests/000-flaky/basic_changelog_changelog-snapshot.t TEST verify_lvm_version; TEST glusterd; TEST pidof glusterd; TEST setup_lvm 1 path: glusterfs_11.2-3/tests/bugs/nfs/bug-970070.t cleanup TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 $H0:$B0/$V0 path: glusterfs_11.2-3/tests/basic/afr/replace-brick-self-heal-non-granular.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/${V0}{0,1} TEST $CLI volume set $V0 cluster.granular-entry-heal off path: glusterfs_11.2-3/tests/000-flaky/bugs_glusterd_bug-857330/normal.t TEST glusterd TEST pidof glusterd TEST $CLI volume info; path: glusterfs_11.2-3/tests/bugs/snapshot/bug-1045333-zfs.t TEST verify_zfs_version; TEST glusterd; TEST pidof glusterd; TEST setup_zfs 1 path: glusterfs_11.2-3/tests/bugs/glusterfs/bug-902610.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 $H0:$B0/${V0}0 $H0:$B0/${V0}1 $H0:$B0/${V0}2 $H0:$B0/${V0}3 path: glusterfs_11.2-3/tests/bugs/shard/bug-1251824.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/${V0}{0,1,2,3} TEST $CLI volume set $V0 features.shard on path: glusterfs_11.2-3/tests/bugs/shard/parallel-truncate-read.t TEST touch /tmp/parallel-truncate-read TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 3 $H0:$B0/${V0}{0,1,2} TEST $CLI volume set $V0 features.shard on path: glusterfs_11.2-3/tests/basic/geo-replication/marker-xattrs.t TEST glusterd TEST pidof glusterd ## Start and create a replicated volume path: glusterfs_11.2-3/tests/basic/geo-replication/marker-xattrs.t cleanup TEST glusterd TEST pidof glusterd ## Start and create a disperse volume TEST $CLI volume create $V0 disperse 3 redundancy 1 $H0:$B0/${V0}-{0,1,2} path: glusterfs_11.2-3/tests/bugs/glusterfs-server/bug-873549.t TEST glusterd -LDEBUG; TEST pidof glusterd; TEST $CLI volume info; path: glusterfs_11.2-3/tests/bugs/distribute/bug-1088231.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 $H0:$B0/${V0}{0,1} TEST $CLI volume set $V0 cluster.randomize-hash-range-by-gfid on path: glusterfs_11.2-3/tests/bugs/distribute/bug-1088231.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 $H0:$B0/${V0}{0,1} TEST $CLI volume set $V0 cluster.randomize-hash-range-by-gfid on path: glusterfs_11.2-3/tests/bugs/nfs/showmount-many-clients.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 $H0:$B0/brick1 path: glusterfs_11.2-3/tests/bugs/shard/issue-1358.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/${V0}{0,1} TEST $CLI volume set $V0 features.shard on path: glusterfs_11.2-3/tests/bugs/shard/issue-1281.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 $H0:$B0/${V0}{0,1} TEST $CLI volume set $V0 features.shard on path: glusterfs_11.2-3/tests/bugs/replicate/bug-1190069-afr-stale-index-entries.t TEST glusterd; TEST pidof glusterd; TEST $CLI volume create $V0 replica 2 $H0:$B0/${V0}{0,1}; TEST $CLI volume set $V0 cluster.self-heal-daemon off path: glusterfs_11.2-3/tests/bugs/nfs/subdir-trailing-slash.t cleanup; TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 $H0:$B0/$V0 path: glusterfs_11.2-3/tests/features/simple-quota.t TEST glusterd TEST pidof glusterd TEST $CLI volume info; path: glusterfs_11.2-3/tests/bugs/gfapi/bug-1447266/bug-1447266.t TEST glusterd; TEST pidof glusterd; TEST $CLI volume create $V0 $H0:$L1 $H0:$L2 $H0:$L3; path: glusterfs_11.2-3/tests/basic/afr/afr-seek.t TEST glusterd TEST pidof glusterd TEST $CLI volume info path: glusterfs_11.2-3/tests/000-flaky/bugs_distribute_bug-1122443.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 $H0:$B0/${V0}0 path: glusterfs_11.2-3/tests/bugs/quick-read/bz1523599/bz1523599.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 $H0:$B0/brick1; path: glusterfs_11.2-3/tests/basic/changelog/changelog-api.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 $H0:$B0/${V0}0 path: glusterfs_11.2-3/tests/basic/graph-cleanup-brick-down-shd-mux.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 3 $H0:$B0/${V0}{0,1,2,3,4,5} TEST $CLI volume set $V0 cluster.background-self-heal-count 0 path: glusterfs_11.2-3/tests/bugs/replicate/bug-1365455.t cleanup; TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 $H0:$B0/${V0}0; path: glusterfs_11.2-3/tests/basic/ctime/ctime-rep-heal.t #cleate and start volume TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 3 $H0:$B0/${V0}{1..3} TEST $CLI volume start $V0 path: glusterfs_11.2-3/tests/bugs/posix/disallow-gfid-volumeid-fremovexattr.t ## Start and create a volume TEST glusterd; TEST pidof glusterd; TEST $CLI volume info; path: glusterfs_11.2-3/tests/basic/volfile-sanity.t ## Start and create a volume TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 $H0:$B0/${V0}1 path: glusterfs_11.2-3/tests/basic/afr/afr-anon-inode.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 3 $H0:$B0/${V0}{0..2} TEST $CLI volume set $V0 performance.quick-read off path: glusterfs_11.2-3/tests/basic/quota-anon-fd-nfs.t TESTS_EXPECTED_IN_LOOP=16 TEST glusterd TEST pidof glusterd TEST $CLI volume info; path: glusterfs_11.2-3/tests/bugs/replicate/bug-1477169-entry-selfheal-rename.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 3 $H0:$B0/${V0}{0,1,2} TEST $CLI volume start $V0 path: glusterfs_11.2-3/tests/basic/stats-dump.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 3 $H0:$B0/${V0}{0,1,2} TEST $CLI volume set $V0 diagnostics.latency-measurement on path: glusterfs_11.2-3/tests/bugs/quick-read/bug-846240.t TEST glusterd TEST pidof glusterd TEST $CLI volume info; path: glusterfs_11.2-3/tests/bugs/glusterfs/bug-856455.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 $H0:$B0/${V0}0 $H0:$B0/${V0}1 $H0:$B0/${V0}2 path: glusterfs_11.2-3/tests/bugs/snapshot/bug-1109770.t TEST glusterd; TEST pidof glusterd; TEST $CLI volume create $V0 $H0:$L1 $H0:$L2 $H0:$L3; path: glusterfs_11.2-3/tests/bugs/replicate/bug-1101647.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/${V0}{1,2}; TEST $CLI volume start $V0; path: glusterfs_11.2-3/tests/bugs/gfapi/bug-1032894.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/${V0}{0,1} TEST $CLI volume set $V0 cluster.self-heal-daemon off path: glusterfs_11.2-3/tests/bugs/glusterd/bug-1595320.t # Start glusterd TEST glusterd TEST pidof glusterd # Create volume and enable brick multiplexing path: glusterfs_11.2-3/tests/bugs/glusterfs-server/bug-864222.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 $H0:$B0/brick0 TEST $CLI volume set $V0 nfs.disable false path: glusterfs_11.2-3/tests/basic/ec/ec-anonymous-fd.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 disperse 6 redundancy 2 $H0:$B0/${V0}{0..5} TEST $CLI volume start $V0 path: glusterfs_11.2-3/tests/bugs/distribute/bug-1600379.t # Start glusterd TEST glusterd; TEST pidof glusterd; TEST $CLI volume info; path: glusterfs_11.2-3/tests/bugs/distribute/bug-924265.t TEST glusterd TEST pidof glusterd mkdir -p $H0:$B0/${V0}0 path: glusterfs_11.2-3/tests/bugs/locks/issue-2443-crash.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 $H0:$B0/brick0 TEST $CLI volume start $V0 path: glusterfs_11.2-3/tests/bugs/distribute/bug-1193636.t #Basic checks TEST glusterd TEST pidof glusterd TEST $CLI volume info path: glusterfs_11.2-3/tests/bugs/distribute/bug-973073.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 $H0:$B0/${V0}0 $H0:$B0/${V0}1 $H0:$B0/${V0}2 path: glusterfs_11.2-3/tests/basic/distribute/file-create.t TEST glusterd TEST pidof glusterd # We want fixed size bricks to test min-free-disk path: glusterfs_11.2-3/tests/bugs/replicate/bug-1408712.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 3 $H0:$B0/${V0}{0,1,2} path: glusterfs_11.2-3/tests/bugs/replicate/bug-821056.t cleanup; TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/${V0}{0,1} path: glusterfs_11.2-3/tests/bugs/shard/bug-1258334.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/${V0}{0,1,2,3} TEST $CLI volume set $V0 features.shard on path: glusterfs_11.2-3/tests/bugs/unclassified/bug-1357397.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/${V0}{1,2} path: glusterfs_11.2-3/tests/bugs/cli/bug-764638.t TEST glusterd TEST pidof glusterd TEST $CLI pool list; path: glusterfs_11.2-3/tests/basic/afr/stale-file-lookup.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/${V0}{0,1} TEST $CLI volume set $V0 self-heal-daemon off path: glusterfs_11.2-3/tests/basic/afr/granular-esh/granular-esh.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/${V0}{0,1} TEST $CLI volume start $V0 path: glusterfs_11.2-3/tests/basic/afr/sparse-file-self-heal.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/${V0}{0,1} TEST $CLI volume set $V0 data-self-heal-algorithm full path: glusterfs_11.2-3/tests/000-flaky/basic_afr_split-brain-favorite-child-policy.t TEST glusterd TEST pidof glusterd #Create replica 2 volume path: glusterfs_11.2-3/tests/bugs/replicate/bug-1756938-replica-3-sbrain-cli.t TEST glusterd; TEST pidof glusterd; TEST $CLI volume create $V0 replica 3 $H0:$B0/${V0}{0,1,2} TEST $CLI volume set $V0 features.shard enable path: glusterfs_11.2-3/tests/basic/afr/granular-esh/cli.t TEST glusterd TEST pidof glusterd ######################### path: glusterfs_11.2-3/tests/bugs/snapshot/bug-1162462.t TEST setup_lvm 3; TEST glusterd; TEST pidof glusterd; TEST $CLI volume create $V0 $H0:$L1 $H0:$L2 $H0:$L3; path: glusterfs_11.2-3/tests/bitrot/bug-1207627-bitrot-scrub-status.t ## Start glusterd TEST glusterd; TEST pidof glusterd; ## Lets create and start the volume path: glusterfs_11.2-3/tests/bugs/cli/bug-961307.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica $REPLICA $H0:$B0/${V0}-00 $H0:$B0/${V0}-01 $H0:$B0/${V0}-10 $H0:$B0/${V0}-11 path: glusterfs_11.2-3/tests/bugs/glusterfs-server/bug-852147.t ## Start and create a volume TEST glusterd; TEST pidof glusterd; TEST $CLI volume info; path: glusterfs_11.2-3/tests/bugs/replicate/bug-884328.t cleanup; TEST glusterd TEST pidof glusterd TEST check_option_help_presence "cluster.quorum-type" path: glusterfs_11.2-3/tests/basic/gfproxy.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 3 $H0:$B0/${V0}{0,1,2} TEST $CLI volume set $V0 config.gfproxyd enable path: glusterfs_11.2-3/tests/bugs/replicate/bug-986905.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/${V0}{0,1} TEST $CLI volume start $V0 path: glusterfs_11.2-3/tests/bugs/replicate/bug-1221481-allow-fops-on-dir-split-brain.t TEST glusterd; TEST pidof glusterd; TEST $CLI volume create $V0 replica 2 $H0:$B0/${V0}{0,1}; TEST $CLI volume set $V0 cluster.self-heal-daemon off path: glusterfs_11.2-3/tests/basic/ec/self-heal-read-write-fail.t TEST glusterd TEST pidof glusterd TEST $CLI volume info path: glusterfs_11.2-3/tests/bugs/nfs/bug-1166862.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 $H0:$B0/brick1 path: glusterfs_11.2-3/tests/basic/afr/split-brain-favorite-child-policy-client-side-healing.t TEST glusterd TEST pidof glusterd count_files () { path: glusterfs_11.2-3/tests/bugs/bug-1368312.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/${V0}{0,1,2,3,4,5} TEST $CLI volume start $V0 path: glusterfs_11.2-3/tests/bugs/read-only/bug-1134822-read-only-default-in-graph.t TEST glusterd TEST pidof glusterd #Create a volume path: glusterfs_11.2-3/tests/bugs/distribute/bug-1786679.t TEST glusterd TEST pidof glusterd BRICK1=$B0/${V0}-0 path: glusterfs_11.2-3/tests/bugs/glusterfs/bug-895235.t cleanup; TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/${V0}{0,1} path: glusterfs_11.2-3/tests/basic/ec/ec-read-mask.t cleanup TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 disperse 6 redundancy 2 $H0:$B0/${V0}{0..5} TEST $CLI volume start $V0 path: glusterfs_11.2-3/tests/bugs/replicate/bug-1180545.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/brick{0,1} path: glusterfs_11.2-3/tests/bugs/glusterd/remove-brick-testcases.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 $H0:$B0/${V0}{1..5} path: glusterfs_11.2-3/tests/basic/afr/gfid-self-heal.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/${V0}{0,1} TEST $CLI volume set $V0 self-heal-daemon off path: glusterfs_11.2-3/tests/bugs/bug-1138841.t TEST glusterd TEST pidof glusterd ## Create a volume and set auth.allow using cidr format ip path: glusterfs_11.2-3/tests/bugs/core/949327.t old_count=$(tmp_file_count); TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/${V0}0 $H0:$B0/${V0}1 TEST $CLI volume start $V0 path: glusterfs_11.2-3/tests/basic/pgfid-feat.t TEST glusterd TEST pidof glusterd TEST $CLI volume info; path: glusterfs_11.2-3/tests/bugs/cli/bug-1030580.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/${V0}0 $H0:$B0/${V0}1 # Increasing the json stats dump time interval, so that it doesn't mess with the test. path: glusterfs_11.2-3/tests/bugs/snapshot/bug-1157991-zfs.t TEST verify_zfs_version; TEST glusterd; TEST pidof glusterd; TEST setup_zfs 1 path: glusterfs_11.2-3/tests/basic/glusterd/check-cloudsync-ancestry.t cleanup; TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 $H0:$B0/b1 $H0:$B0/b2 $H0:$B0/b3 path: glusterfs_11.2-3/tests/features/trash.t # testing glusterd TEST glusterd TEST pidof glusterd TEST $CLI volume info path: glusterfs_11.2-3/tests/bugs/distribute/bug-1099890.t TEST glusterd; TEST pidof glusterd; # Create 2 loop devices, one per brick. path: glusterfs_11.2-3/tests/bitrot/bug-1244613.t TESTS_EXPECTED_IN_LOOP=16 TEST glusterd TEST pidof glusterd TEST $CLI volume info; path: glusterfs_11.2-3/tests/basic/afr/resolve.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/${V0}{0,1} TEST $CLI volume set $V0 cluster.self-heal-daemon off path: glusterfs_11.2-3/tests/basic/afr/replace-brick-self-heal.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/${V0}{0,1} TEST $CLI volume start $V0 path: glusterfs_11.2-3/tests/bugs/nfs/bug-1302948.t cleanup TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 $H0:$B0/$V0 path: glusterfs_11.2-3/tests/bugs/replicate/bug-1686568-send-truncate-on-arbiter-from-shd.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/${V0}{0,1} TEST $CLI volume set $V0 changelog.changelog on path: glusterfs_11.2-3/tests/basic/distribute/dht_seek_test.t # Start glusterd TEST glusterd; TEST pidof glusterd; TEST $CLI volume info; path: glusterfs_11.2-3/tests/line-coverage/some-features-in-libglusterfs.t TEST glusterd TEST pidof glusterd TEST $CLI volume info; path: glusterfs_11.2-3/tests/basic/distribute/manual-rebalance.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 $H0:$B0/${V0}0 TEST $CLI volume set $V0 performance.quick-read off path: glusterfs_11.2-3/tests/bugs/glusterfs/bug-866459.t ## Start and create a volume TEST glusterd TEST pidof glusterd TEST $CLI volume info; path: glusterfs_11.2-3/tests/bugs/bitrot/bug-1245981.t ## Start glusterd TEST glusterd; TEST pidof glusterd; ## Lets create and start the volume path: glusterfs_11.2-3/tests/basic/fop-sampling.t cleanup; TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 3 $H0:$B0/${V0}{0,1,2} $CLI volume set $V0 nfs.disable off path: glusterfs_11.2-3/tests/basic/afr/self-heald.t TESTS_EXPECTED_IN_LOOP=20 TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/${V0}{0,1,2,3,4,5} TEST $CLI volume set $V0 cluster.background-self-heal-count 0 path: glusterfs_11.2-3/tests/basic/ec/ec-common TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 redundancy $REDUNDANCY $H0:$B0/${V0}{0..$LAST_BRICK} EXPECT 'Created' volinfo_field $V0 'Status' path: glusterfs_11.2-3/tests/bugs/rpc/bug-921072.t #1 TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 $H0:$B0/$V0 path: glusterfs_11.2-3/tests/bugs/glusterfs/bug-879490.t ## Start and create a volume TEST glusterd; TEST pidof glusterd; TEST $CLI volume info; path: glusterfs_11.2-3/tests/bugs/glusterfs/bug-879494.t ## Start and create a volume TEST glusterd; TEST pidof glusterd; TEST $CLI volume info; path: glusterfs_11.2-3/tests/bugs/replicate/bug-978794.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/${V0}{0,1} TEST $CLI volume start $V0 path: glusterfs_11.2-3/tests/bugs/bug-1694920.t TEST glusterd; TEST pidof glusterd TEST $CLI volume create $V0 $H0:$B0/${V0}; path: glusterfs_11.2-3/tests/bugs/replicate/bug-1402730.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 3 $H0:$B0/${V0}{0,1,2} TEST $CLI volume set $V0 granular-entry-heal on path: glusterfs_11.2-3/tests/bugs/posix/bug-gfid-path.t TEST glusterd TEST pidof glusterd TEST $CLI volume info; path: glusterfs_11.2-3/tests/bugs/quota/bug-1292020.t TEST glusterd; TEST pidof glusterd; TEST $CLI volume create $V0 $H0:$B0/$V0 path: glusterfs_11.2-3/tests/bugs/shard/unlinks-and-renames.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/${V0}{0,1} TEST $CLI volume set $V0 features.shard on path: glusterfs_11.2-3/tests/bugs/shard/unlinks-and-renames.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/${V0}{0,1} TEST $CLI volume set $V0 features.shard on path: glusterfs_11.2-3/tests/bugs/snapshot/bug-1260848.t ## Start and create a volume TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 disperse 6 redundancy 2 $H0:$B0/${V0}{0..5} path: glusterfs_11.2-3/tests/bugs/replicate/bug-886998.t gfid1="0xc2e75dde97f346e7842d1076a8e699f8" TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/${V0}0 $H0:$B0/${V0}1 TEST $CLI volume start $V0 path: glusterfs_11.2-3/tests/bugs/replicate/bug-1238398-split-brain-resolution.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/${V0}{0,1} TEST $CLI volume start $V0 path: glusterfs_11.2-3/tests/basic/afr/lk-quorum.t TEST glusterd; TEST pidof glusterd #Tests for quorum-type option for replica 2 path: glusterfs_11.2-3/tests/basic/afr/lk-quorum.t cleanup; TEST glusterd; TEST pidof glusterd #Tests for replica 3 path: glusterfs_11.2-3/tests/bugs/glusterfs/bug-861015-index.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/${V0}{0,1,2,3,4,5} TEST $CLI volume set $V0 cluster.ensure-durability off path: glusterfs_11.2-3/tests/bugs/glusterfs/bug-872923.t TEST glusterd TEST pidof glusterd TEST $CLI volume info TEST $CLI volume create $V0 replica 2 $H0:$B0/brick0 $H0:$B0/brick1 path: glusterfs_11.2-3/tests/features/index/index-link-count-lifecycle.t #Same tests for EC volume, EC doesn't fetch link-count, so it is not refreshed TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 disperse 3 $H0:$B0/brick{0,1,2} TEST $CLI volume set $V0 performance.stat-prefetch off path: glusterfs_11.2-3/tests/features/index/index-link-count-lifecycle.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 3 $H0:$B0/brick{0,1,2} TEST $CLI volume set $V0 performance.stat-prefetch off path: glusterfs_11.2-3/tests/features/index/index-link-count-lifecycle.t #Same tests for distribute volume TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 $H0:$B0/brick0 TEST $CLI volume set $V0 performance.stat-prefetch off path: glusterfs_11.2-3/tests/bugs/ec/bug-1227869.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 disperse 3 $H0:$B0/${V0}{1..3} EXPECT "Created" volinfo_field $V0 'Status' path: glusterfs_11.2-3/tests/basic/afr/root-squash-self-heal.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/${V0}{0,1} TEST $CLI volume set $V0 performance.stat-prefetch off path: glusterfs_11.2-3/tests/basic/distribute/debug-xattrs.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 $H0:$B0/$V0-{0..3} path: glusterfs_11.2-3/tests/basic/cdc.t TEST glusterd TEST pidof glusterd ## Create a volume with one brick path: glusterfs_11.2-3/tests/bugs/shard/bug-1342298.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/${V0}{0,1} TEST $CLI volume set $V0 performance.quick-read off path: glusterfs_11.2-3/tests/bugs/shard/bug-1696136-lru-limit-equals-deletion-rate.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 3 $H0:$B0/${V0}{0,1,2} TEST $CLI volume set $V0 features.shard on path: glusterfs_11.2-3/tests/bugs/replicate/bug-977797.t ## Start and create a volume TEST glusterd; TEST pidof glusterd; TEST $CLI volume info; path: glusterfs_11.2-3/tests/glusterfind/glusterfind-basic.t cleanup; TEST glusterd; TEST pidof glusterd ##create .keys path: glusterfs_11.2-3/tests/bugs/posix/disallow-gfid-volumeid-removexattr.t #Basic checks TEST glusterd TEST pidof glusterd TEST $CLI volume info path: glusterfs_11.2-3/tests/basic/glusterd/volume-brick-count.t TEST killall -15 glusterd TEST glusterd TEST pidof glusterd test_volume_config "${V0}_3" "1 x \(2 \+ 1\) = 3" "3" "1" "3" "1" "0" "0" path: glusterfs_11.2-3/tests/basic/glusterd/volume-brick-count.t TESTS_EXPECTED_IN_LOOP=56 TEST glusterd TEST pidof glusterd TEST $CLI volume create ${V0}_1 replica 3 arbiter 1 $H0:$B0/b1 $H0:$B0/b2 $H0:$B0/b3 path: glusterfs_11.2-3/tests/basic/ec/ec-reset-brick.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 disperse 6 redundancy 2 $H0:$B0/${V0}{0..5} TEST $CLI volume start $V0 path: glusterfs_11.2-3/tests/basic/fencing/test-fence-option.t TEST glusterd TEST pidof glusterd TEST $CLI volume info; path: glusterfs_11.2-3/tests/bugs/replicate/bug-1015990-rep.t ## Start and create a volume TEST glusterd; TEST pidof glusterd; TEST $CLI volume info; path: glusterfs_11.2-3/tests/bugs/ec/bug-1547662.t TEST glusterd TEST pidof glusterd TEST ${CLI} volume create ${V0} disperse 6 redundancy 2 ${H0}:${B0}/${V0}{0..5} TEST ${CLI} volume start ${V0} path: glusterfs_11.2-3/tests/basic/afr/entry-self-heal.t zero_xattr="000000000000000000000000" TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/${V0}{0,1} TEST $CLI volume set $V0 self-heal-daemon off path: glusterfs_11.2-3/tests/basic/ec/heal-info.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 redundancy 2 $H0:$B0/${V0}{0..5} TEST $CLI volume set $V0 client-log-level DEBUG path: glusterfs_11.2-3/tests/bugs/shard/bug-1568521-EEXIST.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/${V0}{0,1} TEST $CLI volume set $V0 features.shard on path: glusterfs_11.2-3/tests/bugs/bug-1064147.t # Start glusterd TEST glusterd; TEST pidof glusterd; TEST $CLI volume info; path: glusterfs_11.2-3/tests/basic/afr/client-side-heal.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/${V0}{0,1} TEST $CLI volume set $V0 cluster.self-heal-daemon off path: glusterfs_11.2-3/tests/bugs/shard/bug-1696136.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 3 $H0:$B0/${V0}{0,1,2} TEST $CLI volume set $V0 features.shard on path: glusterfs_11.2-3/tests/bugs/glusterfs-server/bug-877992.t ## Start and create a volume TEST glusterd -LDEBUG TEST pidof glusterd path: glusterfs_11.2-3/tests/bugs/posix/bug-1122028.t TEST glusterd TEST pidof glusterd TEST $CLI volume info path: glusterfs_11.2-3/tests/bugs/glusterd/replace-brick-operations.t # Start glusterd TEST glusterd TEST pidof glusterd ## Lets create and start volume path: glusterfs_11.2-3/tests/basic/distribute/lookup.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 $H0:$B0/$V0-{1..3} path: glusterfs_11.2-3/tests/bugs/changelog/bug-1321955.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/${V0}{0,1} TEST $CLI volume set $V0 changelog.changelog enable path: glusterfs_11.2-3/tests/line-coverage/errorgen-coverage.t ## Start and create a volume TEST glusterd; TEST pidof glusterd; TEST $CLI volume info; path: glusterfs_11.2-3/tests/bugs/posix/bug-1651445.t TEST verify_lvm_version TEST glusterd TEST pidof glusterd TEST init_n_bricks 3 TEST setup_lvm 3 path: glusterfs_11.2-3/tests/000-flaky/basic_mount-nfs-auth.t ## Check whether glusterd is running TEST glusterd TEST pidof glusterd TEST $CLI volume info path: glusterfs_11.2-3/tests/bugs/glusterfs/bug-860297.t TEST glusterd TEST pidof glusterd TEST $CLI volume info TEST $CLI volume create $V0 $H0:$B0/brick1 path: glusterfs_11.2-3/tests/bugs/fuse/bug-1309462.t #Basic checks TEST glusterd TEST pidof glusterd #Create a distributed volume path: glusterfs_11.2-3/tests/bugs/glusterd/optimized-basic-testcases.t TEST glusterd; TEST pidof glusterd; #bug-1238135-lazy-daemon-initialization-on-demand path: glusterfs_11.2-3/tests/bugs/glusterd/optimized-basic-testcases.t function generate_statedump_and_check_for_glusterd_info { pid=`pidof glusterd` #remove old stale statedumps cleanup_statedump $pid path: glusterfs_11.2-3/tests/bugs/glusterd/optimized-basic-testcases.t EXPECT_WITHIN $PROCESS_UP_TIMEOUT "^1$" generate_statedump_and_check_for_glusterd_info cleanup_statedump `pidof glusterd` cleanup path: glusterfs_11.2-3/tests/bugs/glusterd/optimized-basic-testcases.t TEST ! $CLI volume set $V0 performance.open-behind off; TEST pidof glusterd; #bug-1433578 - glusterd should not crash after probing a invalid peer path: glusterfs_11.2-3/tests/bugs/glusterd/optimized-basic-testcases.t TEST ! $CLI peer probe invalid-peer TEST pidof glusterd; TEST $CLI volume create $V0 $H0:$B0/${V0}{1,2}; path: glusterfs_11.2-3/tests/basic/glusterd/disperse-create.t cleanup; TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 disperse $H0:$B0/b1 $H0:$B0/b2 $H0:$B0/b3 EXPECT "1 x \(2 \+ 1\) = 3" volinfo_field $V0 "Number of Bricks" path: glusterfs_11.2-3/tests/bugs/bug-1258069.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 3 $H0:$B0/${V0}{0,1,2} TEST $CLI volume set $V0 cluster.choose-local off path: glusterfs_11.2-3/tests/bugs/snapshot/bug-1166197.t TEST verify_lvm_version; TEST glusterd; TEST pidof glusterd; TEST setup_lvm 1 path: glusterfs_11.2-3/tests/basic/distribute/bug-1265677-use-readdirp.t cleanup TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 $H0:$B0/${V0}{0..1} TEST $CLI volume set $V0 nfs.disable yes path: glusterfs_11.2-3/tests/basic/sdfs-sanity.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 3 $H0:$B0/${V0}{1,2,3,4,5,6,7,8,9}; path: glusterfs_11.2-3/tests/basic/ec/ec.t TEST glusterd TEST pidof glusterd TEST $CLI volume info path: glusterfs_11.2-3/tests/bugs/replicate/bug-1539358-split-brain-detection.t ## Start and create a volume TEST glusterd; TEST pidof glusterd; TEST $CLI volume info; path: glusterfs_11.2-3/tests/00-geo-rep/00-georep-verify-non-root-setup.t cleanup; TEST glusterd; TEST pidof glusterd path: glusterfs_11.2-3/tests/00-geo-rep/00-georep-verify-non-root-setup.t TEST killall_gluster; TEST glusterd; TEST pidof glusterd; ##Create, start and mount meta_volume path: glusterfs_11.2-3/tests/bugs/snapshot/bug-1232430-zfs.t TEST verify_zfs_version; TEST glusterd -LDEBUG; TEST pidof glusterd; TEST setup_zfs 1 path: glusterfs_11.2-3/tests/bugs/glusterfs-server/bug-905864.t ## Start and create a volume TEST glusterd; TEST pidof glusterd; TEST $CLI volume info; path: glusterfs_11.2-3/tests/bugs/snapshot/bug-1113975.t TEST glusterd; TEST pidof glusterd; TEST $CLI volume create $V0 $H0:$L1 $H0:$L2 $H0:$L3; path: glusterfs_11.2-3/tests/bugs/glusterfs/bug-867253.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 $H0:$B0/${V0}0 $H0:$B0/${V0}1 path: glusterfs_11.2-3/tests/bugs/snapshot/bug-1399598-uss-with-ssl.t TEST glusterd TEST pidof glusterd; EXPECT_WITHIN $CHILD_UP_TIMEOUT "1" volume_online_brick_count EXPECT_WITHIN $PROCESS_UP_TIMEOUT "3" total_online_bricks path: glusterfs_11.2-3/tests/bugs/snapshot/bug-1399598-uss-with-ssl.t # Start glusterd TEST glusterd TEST pidof glusterd; #EST $CLI volume set all cluster.brick-multiplex on path: glusterfs_11.2-3/tests/bugs/core/bug-986429.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 $H0:$B0/$V0 path: glusterfs_11.2-3/tests/000-flaky/bugs_glusterd_bug-857330/xml.t TEST glusterd TEST pidof glusterd TEST $CLI volume info; path: glusterfs_11.2-3/tests/basic/afr/bug-1130892-non-granular.t TEST glusterd TEST pidof glusterd TEST $CLI volume info; path: glusterfs_11.2-3/tests/bugs/glusterfs/bug-873962.t TEST glusterd TEST pidof glusterd TEST $CLI volume info; path: glusterfs_11.2-3/tests/bugs/snapshot/bug-1045333.t TEST verify_lvm_version; TEST glusterd; TEST pidof glusterd; TEST setup_lvm 1 path: glusterfs_11.2-3/tests/bugs/distribute/bug-1368012.t ## Start glusterd TEST glusterd; TEST pidof glusterd; TEST $CLI volume info; path: glusterfs_11.2-3/glusterfs.spec.in %endif pidof -c -o %PPID -x glusterd &> /dev/null if [ $? -eq 0 ]; then kill -9 `pgrep -f gsyncd.py` &> /dev/null path: glusterfs_11.2-3/tests/bugs/glusterfs/bug-844688.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 $H0:$B0/brick0 TEST $CLI volume start $V0 path: glusterfs_11.2-3/tests/basic/distribute/force-migration.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 $H0:$B0/${V0}0 $H0:$B0/${V0}1 TEST $CLI volume set $V0 performance.quick-read off path: glusterfs_11.2-3/tests/bugs/ec/bug-1304988.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 disperse 6 redundancy 2 $H0:$B0/${V0}{0..5} TEST $CLI volume start $V0 path: glusterfs_11.2-3/tests/bugs/ctime/issue-832.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 $H0:$B0/$V0 path: glusterfs_11.2-3/tests/000-flaky/features_lock-migration_lkmigration-set-option.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 $H0:$B0/brick1 $H0:$B0/brick2 TEST $CLI volume start $V0 path: glusterfs_11.2-3/tests/bugs/rpc/bug-1043886.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/${V0}{1,2}; TEST $CLI volume set $V0 nfs.disable false path: glusterfs_11.2-3/extras/init.d/glusterd-Redhat.in start() { if pidofproc -p $PIDFILE $GLUSTERD_BIN &> /dev/null; then echo "glusterd service is already running with pid $PID" return 0 path: glusterfs_11.2-3/extras/init.d/glusterd-Redhat.in { echo -n $"Stopping $BASE:" if pidofproc -p $PIDFILE $GLUSTERD_BIN &> /dev/null; then killproc -p $PIDFILE $BASE else path: glusterfs_11.2-3/tests/bugs/snapshot/bug-1322772-real-path-fix-for-snapshot.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 $H0:$L1 path: glusterfs_11.2-3/tests/bugs/snapshot/bug-1322772-real-path-fix-for-snapshot.t TEST glusterd TEST pidof glusterd EXPECT_WITHIN $PROCESS_UP_TIMEOUT "1" brick_up_status $V0 $H0 $L1 path: glusterfs_11.2-3/tests/bugs/core/bug-1119582.t TEST glusterd; TEST pidof glusterd; TEST $CLI volume create $V0 $H0:$B0/${V0}0 $H0:$B0/${V0}1 path: glusterfs_11.2-3/tests/bugs/replicate/bug-1139230.t TEST glusterd TEST pidof glusterd TEST $CLI volume info; path: glusterfs_11.2-3/tests/bugs/protocol/bug-808400-dist.t TEST glusterd TEST pidof glusterd TEST $CLI volume info; path: glusterfs_11.2-3/tests/basic/afr/metadata-self-heal.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/brick{0,1} TEST $CLI volume start $V0 path: glusterfs_11.2-3/tests/bugs/snapshot/bug-1162498.t TEST verify_lvm_version; TEST glusterd; TEST pidof glusterd; TEST setup_lvm 1 path: glusterfs_11.2-3/tests/basic/distribute/migration-fsync.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 $H0:$B0/${V0}0 TEST $CLI volume set $V0 performance.quick-read off path: glusterfs_11.2-3/tests/bugs/posix/bug-1360679.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 disperse 6 redundancy 2 $H0:$B0/${V0}{0..5} TEST $CLI volume start $V0 path: glusterfs_11.2-3/tests/basic/afr/quorum.t cleanup; TEST glusterd; TEST pidof glusterd #Tests for quorum-type option for replica 3 path: glusterfs_11.2-3/tests/basic/afr/quorum.t TEST glusterd; TEST pidof glusterd function test_write { path: glusterfs_11.2-3/tests/basic/ec/ec-badfd.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 disperse 6 redundancy 2 $H0:$B0/${V0}{1..6} path: glusterfs_11.2-3/tests/bugs/distribute/bug-853258.t TEST glusterd TEST pidof glusterd mkdir -p $H0:$B0/${V0}0 path: glusterfs_11.2-3/tests/gfid2path/gfid2path_fuse.t TEST [ -x $XXHSUM_EXEC ] TEST glusterd TEST pidof glusterd ## Create a single brick volume (B=1) path: glusterfs_11.2-3/tests/basic/afr/entry-self-heal-anon-dir-off.t zero_xattr="000000000000000000000000" TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/${V0}{0,1} TEST $CLI volume heal $V0 disable path: glusterfs_11.2-3/tests/bugs/locks/issue-2551.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 disperse 3 redundancy 1 $H0:$B0/brick{0..2} TEST $CLI volume set $V0 disperse.eager-lock on path: glusterfs_11.2-3/tests/basic/ec/ec-readdir.t cleanup TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 disperse 3 redundancy 1 $H0:$B0/${V0}{0..5} TEST $CLI volume heal $V0 disable path: glusterfs_11.2-3/tests/bugs/replicate/bug-1341650.t TEST glusterd TEST pidof glusterd TEST $CLI volume info path: glusterfs_11.2-3/tests/bugs/glusterfs/bug-906646.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica $REPLICA $H0:$B0/${V0}-00 $H0:$B0/${V0}-01 $H0:$B0/${V0}-10 $H0:$B0/${V0}-11 path: glusterfs_11.2-3/tests/bugs/fuse/bug-924726.t TEST glusterd TEST pidof glusterd TEST $CLI volume info; path: glusterfs_11.2-3/tests/basic/afr/add-brick-self-heal.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/${V0}{0,1} EXPECT 'Created' volinfo_field $V0 'Status'; path: glusterfs_11.2-3/tests/basic/ec/ec-cpu-extensions.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 redundancy $REDUNDANCY $H0:$B0/${V0}{1..$DISPERSE} TEST $CLI volume set $V0 performance.flush-behind off path: glusterfs_11.2-3/tests/00-geo-rep/georep-config-upgrade.t cleanup; TEST glusterd; TEST pidof glusterd ##Variables path: glusterfs_11.2-3/tests/bitrot/br-state-check.t TEST glusterd TEST pidof glusterd ## Create a distribute volume path: glusterfs_11.2-3/tests/bugs/ec/bug-1251446.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 disperse 4 redundancy 1 $H0:$B0/${V0}{0..3} TEST $CLI volume start $V0 path: glusterfs_11.2-3/tests/bugs/glusterd/issue-3781.t TEST glusterd TEST pidof glusterd TEST ${CLI} volume create ${V0} ${H0}:${B0}/brick path: glusterfs_11.2-3/tests/basic/afr/add-brick-self-heal-non-granular.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/${V0}{0,1} EXPECT 'Created' volinfo_field $V0 'Status'; path: glusterfs_11.2-3/tests/bugs/cli/bug-969193.t TEST glusterd TEST pidof glusterd TEST $CLI volume info TEST $CLI volume create $V0 $H0:$B0/brick1 path: glusterfs_11.2-3/tests/bugs/replicate/bug-1744548-heal-timeout.t TEST glusterd; TEST pidof glusterd; TEST $CLI volume create $V0 replica 3 $H0:$B0/${V0}{0,1,2} TEST $CLI volume heal $V0 disable path: glusterfs_11.2-3/tests/basic/afr/bug-1749322-entry-heal-not-happening-non-granular.t ## Start and create a volume TEST glusterd; TEST pidof glusterd; TEST $CLI volume info; path: glusterfs_11.2-3/tests/00-geo-rep/01-georep-glusterd-tests.t TEST pkill glusterd TEST glusterd; TEST pidof glusterd #Start geo-rep and validate secondary format is updated path: glusterfs_11.2-3/tests/00-geo-rep/01-georep-glusterd-tests.t cleanup; TEST glusterd; TEST pidof glusterd path: glusterfs_11.2-3/tests/basic/gfapi/glfs_sysrq.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 $H0:$B0/brick1 EXPECT 'Created' volinfo_field $V0 'Status' path: glusterfs_11.2-3/tests/gfid2path/gfid2path_nfs.t TEST [ -x $XXHSUM_EXEC ] TEST glusterd TEST pidof glusterd ## Create a single brick volume (B=1) path: glusterfs_11.2-3/tests/basic/ec/ec-read-policy.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 disperse 6 redundancy 2 $H0:$B0/${V0}{0..5} TEST $CLI volume set $V0 performance.quick-read off path: glusterfs_11.2-3/tests/bugs/replicate/bug-1655854-support-dist-to-rep3-arb-conversion.t TEST glusterd TEST pidof glusterd # Conversion from 2x1 to 2x3 path: glusterfs_11.2-3/tests/bugs/replicate/bug-1561129-enospc.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 3 $H0:$B0/bricks/brick{0,1,3} TEST $CLI volume start $V0 path: glusterfs_11.2-3/tests/basic/ec/ec-rebalance.t #cleate and start volume TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 disperse 3 redundancy 1 $H0:$B0/${V0}{0..2} TEST $CLI volume set $V0 lookup-optimize on path: glusterfs_11.2-3/tests/bugs/cli/bug-1047416.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/${V0}0 $H0:$B0/${V0}1 # Increasing the json stats dump time interval, so that it doesn't mess with the test. path: glusterfs_11.2-3/tests/basic/multiple-volume-shd-mux.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 3 $H0:$B0/${V0}{0,1,2,3,4,5} TEST $CLI volume start $V0 path: glusterfs_11.2-3/tests/bugs/shard/bug-1568521.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/${V0}{0,1} TEST $CLI volume set $V0 features.shard on path: glusterfs_11.2-3/tests/bugs/transport/bug-873367.t TEST glusterd TEST pidof glusterd TEST $CLI volume info; path: glusterfs_11.2-3/tests/bugs/cli/bug-949298.t cleanup; TEST glusterd TEST pidof glusterd TEST $CLI --xml volume info $V0 path: glusterfs_11.2-3/tests/basic/fencing/afr-lock-heal-advanced.t TEST glusterd TEST pidof glusterd TEST $CLI volume info; path: glusterfs_11.2-3/tests/bugs/distribute/bug-1125824.t TEST glusterd TEST pidof glusterd TEST $CLI volume info; path: glusterfs_11.2-3/tests/basic/ctime/ctime-noatime.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 3 $H0:$B0/${V0}{0,1,2} TEST $CLI volume set $V0 performance.stat-prefetch off path: glusterfs_11.2-3/tests/bugs/distribute/bug-912564.t TEST glusterd TEST pidof glusterd mkdir -p $H0:$B0/${V0}0 path: glusterfs_11.2-3/tests/000-flaky/glusterd-restart-shd-mux.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 3 $H0:$B0/${V0}{0,1,2,3,4,5} TEST $CLI volume set $V0 cluster.background-self-heal-count 0 path: glusterfs_11.2-3/tests/bugs/replicate/886998/strict-readdir.t cleanup; TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/r2d2_0 $H0:$B0/r2d2_1 $H0:$B0/r2d2_2 $H0:$B0/r2d2_3 path: glusterfs_11.2-3/tests/basic/distribute/dir-heal.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 $H0:$B0/$V0-{1..3} path: glusterfs_11.2-3/tests/bugs/replicate/bug-1655050-dir-sbrain-size-policy.t TEST glusterd TEST pidof glusterd #Create replica 2 volume path: glusterfs_11.2-3/tests/bugs/upcall/bug-1227204.t TEST glusterd TEST pidof glusterd; TEST $CLI volume info; path: glusterfs_11.2-3/tests/bugs/snapshot/bug-1162462-zfs.t TEST setup_zfs 3; TEST glusterd; TEST pidof glusterd; TEST $CLI volume create $V0 $H0:$L1 $H0:$L2 $H0:$L3; path: glusterfs_11.2-3/tests/bugs/rpc/bug-884452.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 $H0:$B0/$V0 path: glusterfs_11.2-3/tests/bugs/glusterfs/bug-861015-log.t log_wd=$(gluster --print-logdir) TEST glusterd TEST pidof glusterd rm -f $log_wd/glustershd.log TEST $CLI volume create $V0 replica 2 $H0:$B0/${V0}{0,1} path: glusterfs_11.2-3/tests/bugs/fuse/bug-1126048.t } TEST glusterd TEST pidof glusterd TEST $CLI volume info; path: glusterfs_11.2-3/tests/bugs/distribute/bug-961615.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 $H0:$B0/${V0}0 $H0:$B0/${V0}1 TEST $CLI volume set $V0 performance.quick-read off path: glusterfs_11.2-3/tests/bugs/glusterfs/bug-848251.t TEST glusterd TEST pidof glusterd TEST $CLI volume info; path: glusterfs_11.2-3/tests/bugs/snapshot/bug-1279327.t TEST verify_lvm_version; TEST glusterd; TEST pidof glusterd; TEST init_n_bricks 3 path: glusterfs_11.2-3/tests/basic/afr/bug-1493415-gfid-heal-non-granular.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/${V0}{0,1} TEST $CLI volume set $V0 cluster.granular-entry-heal off path: glusterfs_11.2-3/tests/basic/ec/ec-quorum-count.t cleanup TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 disperse 6 redundancy 2 $H0:$B0/${V0}{0..5} TEST $CLI volume create $V1 $H0:$B0/${V1}{0..5} path: glusterfs_11.2-3/tests/include.rc } which pidof > /dev/null || { pidof() { $PYTHON pidof.py $@ path: glusterfs_11.2-3/tests/include.rc which pidof > /dev/null || { pidof() { $PYTHON pidof.py $@ } } path: glusterfs_11.2-3/tests/include.rc which pidof > /dev/null || { pidof() { $PYTHON pidof.py $@ } path: glusterfs_11.2-3/tests/basic/volume-scale-shd-mux.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 3 $H0:$B0/${V0}{0,1,2,3,4,5} TEST $CLI volume set $V0 cluster.background-self-heal-count 0 path: glusterfs_11.2-3/tests/basic/volume-snapshot-xml.t TEST verify_lvm_version; TEST glusterd; TEST pidof glusterd; TEST setup_lvm 1 path: glusterfs_11.2-3/tests/bugs/snapshot/bug-1399598-uss-with-ssl-zfs.t # Start glusterd TEST glusterd TEST pidof glusterd; #EST $CLI volume set all cluster.brick-multiplex on path: glusterfs_11.2-3/tests/bugs/snapshot/bug-1399598-uss-with-ssl-zfs.t TEST glusterd TEST pidof glusterd; EXPECT_WITHIN $CHILD_UP_TIMEOUT "1" volume_online_brick_count EXPECT_WITHIN $PROCESS_UP_TIMEOUT "3" total_online_bricks path: glusterfs_11.2-3/tests/bugs/shard/issue-1425.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 $H0:$B0/${V0}0 TEST $CLI volume set $V0 features.shard on path: glusterfs_11.2-3/tests/bugs/snapshot/bug-1227646.t TEST glusterd; TEST pidof glusterd; #TEST $CLI volume create $V0 $H0:$L1 $H0:$L2 $H0:$L3; path: glusterfs_11.2-3/tests/bugs/replicate/bug-1238508-self-heal.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/${V0}{0,1} TEST $CLI volume start $V0 path: glusterfs_11.2-3/tests/bugs/distribute/bug-921408.t } TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 $H0:$B0/${V0}0 path: glusterfs_11.2-3/tests/bugs/replicate/bug-767585-gfid.t cleanup; TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/${V0}0 $H0:$B0/${V0}1 TEST $CLI volume start $V0 path: glusterfs_11.2-3/tests/basic/ec/dht-rename.t cleanup TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 disperse 3 redundancy 1 $H0:$B0/${V0}{0..5} TEST $CLI volume start $V0 path: glusterfs_11.2-3/tests/basic/inode-leak.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/${V0}{0,1,2,3} TEST $CLI volume start $V0 path: glusterfs_11.2-3/tests/bugs/nfs/socket-as-fifo.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 $H0:$B0/$V0 path: glusterfs_11.2-3/tests/bugs/replicate/bug-1448804-check-quorum-type-values.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/${V0}{0..1} TEST $CLI volume start $V0 path: glusterfs_11.2-3/tests/basic/distribute/brick-down.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 $H0:$B0/$V0-{1..3} path: glusterfs_11.2-3/tests/bugs/snapshot/bug-1275616-zfs.t TEST verify_zfs_version; TEST glusterd; TEST pidof glusterd; TEST setup_zfs 1 path: glusterfs_11.2-3/tests/bugs/replicate/bug-1250170-fsync.t TEST gcc $(dirname $0)/bug-1250170-fsync.c -o $(dirname $0)/bug-1250170-fsync TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/${V0}{0,1} path: glusterfs_11.2-3/tests/bugs/snapshot/bug-1597662-zfs.t TEST setup_zfs 3; TEST glusterd; TEST pidof glusterd; TEST $CLI volume create $V0 $H0:$L1 $H0:$L2 $H0:$L3; path: glusterfs_11.2-3/tests/features/lock_revocation.t #Init TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/brick{0,1} TEST $CLI volume set $V0 self-heal-daemon off path: glusterfs_11.2-3/tests/bugs/distribute/bug-1161311.t TEST glusterd TEST pidof glusterd TEST $CLI volume info; path: glusterfs_11.2-3/tests/basic/afr/tarissue.t #Basic checks TEST glusterd TEST pidof glusterd #Create a distributed-replicate volume path: glusterfs_11.2-3/tests/basic/afr/arbiter.t TEST glusterd; TEST pidof glusterd # Non arbiter replica 3 volumes should not have arbiter-count option enabled. path: glusterfs_11.2-3/tests/basic/afr/arbiter.t cleanup; TEST glusterd; TEST pidof glusterd # Create and mount a replica 3 arbiter volume. path: glusterfs_11.2-3/tests/bugs/core/bug-1111557.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 $H0:$B0/${V0}{0} TEST $CLI volume set $V0 diagnostics.brick-log-buf-size 0 path: glusterfs_11.2-3/tests/bugs/bitrot/1209751-bitrot-scrub-tunable-reset.t ## Start glusterd TEST glusterd; TEST pidof glusterd; ## Lets create and start the volume path: glusterfs_11.2-3/tests/bugs/bitrot/1209751-bitrot-scrub-tunable-reset.t pkill glusterd; TEST glusterd; TEST pidof glusterd; ## All the bitrot scrub tunable value should come back again. path: glusterfs_11.2-3/tests/bugs/replicate/bug-1046624.t cleanup; TEST glusterd TEST pidof glusterd ## Start and create a volume path: glusterfs_11.2-3/tests/bugs/replicate/bug-916226.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/${V0}0 $H0:$B0/${V0}1 $H0:$B0/${V0}2 $H0:$B0/${V0}3 path: glusterfs_11.2-3/tests/basic/volume-snapshot-xml-zfs.t TEST verify_zfs_version; TEST glusterd; TEST pidof glusterd; TEST setup_zfs 1 path: glusterfs_11.2-3/tests/basic/ec/self-heal.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 redundancy 2 $H0:$B0/${V0}{0..5} TEST $CLI volume set $V0 client-log-level DEBUG path: glusterfs_11.2-3/tests/basic/afr/granular-esh/replace-brick.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/${V0}{0,1} TEST $CLI volume start $V0 path: glusterfs_11.2-3/tests/bugs/snapshot/bug-1168875-zfs.t TEST glusterd; TEST pidof glusterd; TEST $CLI volume create $V0 $H0:$L1 $H0:$L2 $H0:$L3; path: glusterfs_11.2-3/tests/bugs/glusterd/bug-1696046.t ## Start and create a volume TEST glusterd; TEST pidof glusterd; TEST $CLI volume set all cluster.brick-multiplex on path: glusterfs_11.2-3/tests/bugs/shard/bug-1669077.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 3 $H0:$B0/${V0}{0,1,2} TEST $CLI volume set $V0 features.shard on path: glusterfs_11.2-3/tests/00-geo-rep/georep-stderr-hang.t cleanup; TEST glusterd; TEST pidof glusterd path: glusterfs_11.2-3/tests/bugs/shard/bug-1705884.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 3 $H0:$B0/${V0}{0,1,2} TEST $CLI volume set $V0 features.shard on path: glusterfs_11.2-3/tests/bugs/posix/bug-765380.t TEST glusterd TEST pidof glusterd REPLICA=2 path: glusterfs_11.2-3/tests/bugs/posix/posix-reconfigure-uring-crash.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 3 $H0:$B0/${V0}{0..2}; TEST $CLI volume start $V0 path: glusterfs_11.2-3/tests/features/ssl-ciphers.t TEST glusterd TEST pidof glusterd TEST $CLI volume info; path: glusterfs_11.2-3/tests/performance/open-behind.t TEST glusterd TEST pidof glusterd TEST $CLI volume info 2>/dev/null; path: glusterfs_11.2-3/tests/basic/afr/arbiter-cli.t TEST glusterd; TEST pidof glusterd # No replica count. path: glusterfs_11.2-3/tests/basic/ctime/ctime-mdata-legacy-files.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 3 $H0:$B0/${V0}{0,1,2} TEST $CLI volume set $V0 performance.stat-prefetch off path: glusterfs_11.2-3/tests/bugs/replicate/bug-1626994-info-split-brain.t TEST glusterd; TEST pidof glusterd; TEST $CLI volume create $V0 replica 3 $H0:$B0/${V0}{0,1,2}; TEST $CLI volume set $V0 self-heal-daemon off path: glusterfs_11.2-3/tests/bugs/snapshot/bug-1111041.t TEST glusterd; TEST pidof glusterd; TEST $CLI volume create $V0 $H0:$B0/${V0}0 $H0:$B0/${V0}1 path: glusterfs_11.2-3/tests/bugs/glusterd/rebalance-operations-in-single-node.t volname="StartMigrationDuringRebalanceTest" TEST glusterd TEST pidof glusterd; TEST $CLI volume info; path: glusterfs_11.2-3/tests/bugs/glusterd/rebalance-operations-in-single-node.t TEST pkill gluster TEST glusterd TEST pidof glusterd # status should be "completed" immediate after glusterd has respawned. path: glusterfs_11.2-3/tests/bugs/snapshot/bug-1166197-zfs.t TEST verify_zfs_version; TEST glusterd; TEST pidof glusterd; TEST setup_zfs 1 path: glusterfs_11.2-3/tests/bugs/bitrot/bug-1210684-scrub-pause-resume-error-handling.t ## Start glusterd TEST glusterd; TEST pidof glusterd; ## Lets create and start the volume path: glusterfs_11.2-3/tests/bugs/nfs/bug-1161092-nfs-acls.t TEST glusterd TEST pidof glusterd TEST $CLI volume info path: glusterfs_11.2-3/tests/bugs/cli/bug-982174.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 $H0:$B0/brick1 $H0:$B0/brick2 TEST $CLI volume start $V0 path: glusterfs_11.2-3/tests/basic/xlator-pass-through-sanity.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 $H0:$B0/${V0} path: glusterfs_11.2-3/tests/bugs/rpc/bug-847624.t #1 TEST glusterd TEST pidof glusterd #3 TEST $CLI volume create $V0 $H0:$B0/$V0 path: glusterfs_11.2-3/tests/bugs/nfs/bug-1157223-symlink-mounting.t ## Start and create a volume TEST glusterd TEST pidof glusterd TEST $CLI volume info; path: glusterfs_11.2-3/tests/bugs/geo-replication/bug-1296496.t TEST glusterd TEST pidof glusterd ## Start and create a replicated volume path: glusterfs_11.2-3/tests/bugs/replicate/bug-1132102.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/${V0}{0,1} TEST $CLI volume start $V0 path: glusterfs_11.2-3/tests/bugs/glusterfs/bug-1482528.t #Basic checks TEST glusterd TEST pidof glusterd #Create a distributed volume path: glusterfs_11.2-3/tests/bugs/gfapi/glfs_vol_set_IO_ERR.t ## Start and create a volume TEST glusterd; TEST pidof glusterd; TEST $CLI volume info; path: glusterfs_11.2-3/tests/bitrot/bug-1373520.t TEST glusterd TEST pidof glusterd #Create a disperse volume path: glusterfs_11.2-3/tests/bugs/error-gen/bug-767095.t ## Start and create a volume TEST glusterd; TEST pidof glusterd; TEST $CLI volume info; path: glusterfs_11.2-3/tests/bugs/glusterd/add-brick-and-validate-replicated-volume-options.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/${V0}{1,2}; TEST $CLI volume start $V0 path: glusterfs_11.2-3/tests/basic/afr/gfid-mismatch.t #Init TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/brick{0,1} TEST $CLI volume set $V0 self-heal-daemon off path: glusterfs_11.2-3/tests/bugs/access-control/bug-1387241.t ## Start and create a volume TEST glusterd; TEST pidof glusterd; TEST $CLI volume info; path: glusterfs_11.2-3/tests/bugs/replicate/bug-957877.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/${V0}{0,1} TEST $CLI volume start $V0; path: glusterfs_11.2-3/tests/features/flock_interrupt.t ## Start and create a volume TEST glusterd; TEST pidof glusterd; TEST $CLI volume create $V0 $H0:$B0/${V0}0; path: glusterfs_11.2-3/tests/bugs/nfs/bug-877885.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/brick0 $H0:$B0/brick1 TEST $CLI volume set $V0 nfs.disable false path: glusterfs_11.2-3/tests/bugs/shard/bug-shard-discard.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/${V0}{0..3} TEST $CLI volume set $V0 features.shard on path: glusterfs_11.2-3/tests/bugs/bug-1620580.t ## Start glusterd TEST glusterd; TEST pidof glusterd; TEST $CLI volume info; path: glusterfs_11.2-3/tests/bugs/glusterfs-server/bug-887145.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/${V0}{1,2}; TEST $CLI volume set $V0 performance.open-behind off; path: glusterfs_11.2-3/tests/bugs/changelog/bug-1208470.t ## Start and create a volume TEST glusterd; TEST pidof glusterd; TEST $CLI volume create $V0 $H0:$B0/$V0"1"; path: glusterfs_11.2-3/tests/bugs/bug-1110262.t ## Start glusterd TEST glusterd; TEST pidof glusterd; TEST $CLI volume info; path: glusterfs_11.2-3/tests/basic/afr/granular-esh/add-brick.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/${V0}{0,1} TEST $CLI volume start $V0 path: glusterfs_11.2-3/tests/bugs/glusterfs/bug-869724.t ## Start and create a volume TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 $H0:$B0/${V0}1; path: glusterfs_11.2-3/tests/bugs/cli/bug-977246.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 $H0:$B0/${V0}1 path: glusterfs_11.2-3/tests/basic/ec/ec-fast-fgetxattr.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 disperse 6 redundancy 2 $H0:$B0/${V0}{1..6} path: glusterfs_11.2-3/tests/bugs/shard/bug-1605056.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 3 $H0:$B0/${V0}{0,1,2} TEST $CLI volume set $V0 features.shard on path: glusterfs_11.2-3/tests/bugs/shard/configure-lru-limit.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 3 $H0:$B0/${V0}{0,1,2} TEST $CLI volume set $V0 features.shard on path: glusterfs_11.2-3/tests/bugs/glusterd/bug-1242875-do-not-pass-volinfo-quota.t ## Start glusterd TEST glusterd; TEST pidof glusterd; ## Lets create volume V0 and start the volume path: glusterfs_11.2-3/tests/bugs/glusterd/bug-1242875-do-not-pass-volinfo-quota.t ## Start glusterd TEST glusterd; TEST pidof glusterd; ## Quotad daemon should start on restarting the glusterd path: glusterfs_11.2-3/tests/bugs/glusterfs-server/bug-912297.t ## Start and create a volume TEST glusterd; TEST pidof glusterd; TEST $CLI volume create $V0 replica 3 $H0:$B0/${V0}{1,2,3,4,5,6}; path: glusterfs_11.2-3/tests/basic/ec/ec-new-entry.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 disperse 6 redundancy 2 $H0:$B0/${V0}{0..5} TEST $CLI volume start $V0 path: glusterfs_11.2-3/tests/basic/rpc-coverage.t TEST glusterd TEST pidof glusterd TEST $CLI volume info; path: glusterfs_11.2-3/tests/bugs/ec/bug-1161621.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 disperse 5 redundancy 2 $H0:$B0/${V0}{0..4} EXPECT "Created" volinfo_field $V0 'Status' path: glusterfs_11.2-3/tests/bugs/fuse/bug-1030208.t #Basic checks TEST glusterd TEST pidof glusterd TEST $CLI volume info path: glusterfs_11.2-3/tests/bugs/replicate/bug-1696599-io-hang.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 3 $H0:$B0/${V0}{1..3} path: glusterfs_11.2-3/tests/bugs/cli/bug-867252.t TEST glusterd TEST pidof glusterd TEST $CLI volume info; path: glusterfs_11.2-3/tests/basic/ec/ec-internal-xattrs.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 redundancy 2 $H0:$B0/${V0}{0..5} TEST $CLI volume start $V0 path: glusterfs_11.2-3/tests/bugs/distribute/bug-1667804.t #Basic checks TEST glusterd TEST pidof glusterd TEST $CLI volume info path: glusterfs_11.2-3/tests/bugs/replicate/bug-1015990.t ## Start and create a volume TEST glusterd; TEST pidof glusterd; TEST $CLI volume info; path: glusterfs_11.2-3/tests/bugs/replicate/issue-3288-dir-gfid-mismatch-heal-option.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/${V0}{0,1} TEST $CLI volume start $V0 path: glusterfs_11.2-3/tests/basic/quota-ancestry-building.t TESTS_EXPECTED_IN_LOOP=8 TEST glusterd TEST pidof glusterd TEST $CLI volume info; path: glusterfs_11.2-3/tests/bugs/replicate/issue-4064-softlink-entry-recreation.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 3 $H0:$B0/${V0}{0,1,2}; TEST $CLI volume set $V0 cluster.self-heal-daemon off path: glusterfs_11.2-3/tests/bugs/distribute/disallow-fixlayout-on-files.t TEST glusterd TEST pidof glusterd TEST $CLI volume info path: glusterfs_11.2-3/tests/basic/quota_aux_mount.t TEST glusterd TEST pidof glusterd TEST $CLI volume info; path: glusterfs_11.2-3/tests/bugs/replicate/bug-1433571-undo-pending-only-on-up-bricks.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 3 $H0:$B0/${V0}{0,1,2} TEST $CLI volume start $V0 path: glusterfs_11.2-3/tests/bugs/distribute/bug-884455.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 $H0:$B0/${V0}0 $H0:$B0/${V0}1 path: glusterfs_11.2-3/tests/bugs/nfs/bug-904065.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 $H0:$B0/brick1 path: glusterfs_11.2-3/tests/bugs/replicate/bug-1134691-afr-lookup-metadata-heal.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 3 $H0:$B0/brick{0,1,2} TEST $CLI volume set $V0 performance.stat-prefetch off path: glusterfs_11.2-3/tests/bugs/unclassified/bug-991622.t cleanup; TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 $H0:$B0/$V0 path: glusterfs_11.2-3/tests/basic/uss.t TEST glusterd; TEST pidof glusterd; TEST $CLI volume create $V0 $H0:$L1 $H0:$L2 $H0:$L3; path: glusterfs_11.2-3/tests/bugs/shard/bug-1250855.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 $H0:$B0/${V0}{0,1} TEST $CLI volume start $V0 path: glusterfs_11.2-3/tests/bugs/replicate/bug-1032927.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/${V0}{0,1} TEST $CLI volume start $V0 path: glusterfs_11.2-3/tests/bugs/core/bug-913544.t #simulate a split-brain of a file and do truncate. This should not crash the mount point TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/${V0}0 $H0:$B0/${V0}1 TEST $CLI volume set $V0 stat-prefetch off path: glusterfs_11.2-3/tests/bitrot/bug-internal-xattrs-check-1243391.t TEST glusterd TEST pidof glusterd ## Create a distribute volume (B=2) path: glusterfs_11.2-3/tests/basic/distribute/file-rename.t TEST glusterd TEST pidof glusterd path: glusterfs_11.2-3/tests/bugs/io-cache/bug-read-hang.t #Basic checks TEST glusterd TEST pidof glusterd TEST $CLI volume info path: glusterfs_11.2-3/tests/bugs/shard/shard-inode-refcount-test.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 $H0:$B0/${V0}0 TEST $CLI volume set $V0 features.shard on path: glusterfs_11.2-3/tests/basic/ec/ec-root-heal.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 disperse 6 redundancy 2 $H0:$B0/${V0}{0..5} TEST $CLI volume start $V0 path: glusterfs_11.2-3/tests/bugs/replicate/bug-979365.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/${V0}{0,1} TEST $CLI volume set $V0 cluster.ensure-durability on path: glusterfs_11.2-3/tests/bugs/changelog/bug-1225542.t TEST verify_lvm_version TEST glusterd; TEST pidof glusterd; TEST setup_lvm 1 path: glusterfs_11.2-3/tests/basic/mount.t ## Start and create a volume TEST glusterd TEST pidof glusterd TEST $CLI volume info; path: glusterfs_11.2-3/tests/bugs/changelog/bug-1211327.t TEST glusterd; TEST pidof glusterd; ##Let the brick processes starts path: glusterfs_11.2-3/tests/bugs/changelog/bug-1211327.t ## Start and create a volume TEST glusterd; TEST pidof glusterd; TEST $CLI volume create $V0 $H0:$B0/$V0"1"; path: glusterfs_11.2-3/tests/basic/afr/durability-off.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/brick{0,1} TEST $CLI volume start $V0 path: glusterfs_11.2-3/tests/line-coverage/afr-heal-info.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/brick{0,1} TEST $CLI volume set $V0 locking-scheme granular path: glusterfs_11.2-3/tests/bugs/replicate/issue-2232.t TEST gcc $(dirname "${0}")/issue-2232.c -o $(dirname "${0}")/issue-2232 -lgfapi TEST glusterd TEST pidof glusterd TEST $CLI volume create ${V0} replica 3 ${H0}:${B0}/${V0}{0..2} path: glusterfs_11.2-3/tests/bugs/bug-1702299.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 $H0:$B0/${V0}{0,1,2,3} TEST $CLI volume start $V0 path: glusterfs_11.2-3/tests/bugs/access-control/bug-1051896.t ## Start and create a volume TEST glusterd; TEST pidof glusterd; TEST $CLI volume info; path: glusterfs_11.2-3/tests/bugs/fuse/bug-858215.t ## Start and create a volume TEST glusterd; TEST pidof glusterd; TEST $CLI volume info; path: glusterfs_11.2-3/tests/bugs/replicate/bug-830665.t TEST glusterd TEST pidof glusterd TEST $CLI volume info; path: glusterfs_11.2-3/tests/bugs/nfs/bug-1143880-fix-gNFSd-auth-crash.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/${V0}{1,2} TEST $CLI volume set $V0 nfs.disable false path: glusterfs_11.2-3/tests/bugs/replicate/bug-1591193-assign-gfid-and-heal.t # Test on 1x3 volume TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 3 $H0:$B0/${V0}{0,1,2}; TEST $CLI volume start $V0; path: glusterfs_11.2-3/tests/bugs/replicate/bug-1591193-assign-gfid-and-heal.t # Test on 1x (2+1) volume TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 3 arbiter 1 $H0:$B0/${V0}{0,1,2}; TEST $CLI volume start $V0; path: glusterfs_11.2-3/tests/bugs/glusterd/reset-rebalance-state.t TEST glusterd; TEST pidof glusterd; TEST $CLI volume info; path: glusterfs_11.2-3/tests/bugs/core/io-stats-1322825.t TEST glusterd TEST pidof glusterd # Covering replication and distribution in the test path: glusterfs_11.2-3/tests/bugs/distribute/missing-files-rebalance.t TEST glusterd TEST pidof glusterd TEST $CLI volume info path: glusterfs_11.2-3/tests/features/interrupt.t ## Start and create a volume TEST glusterd; TEST pidof glusterd; TEST $CLI volume create $V0 replica 3 $H0:$B0/${V0}{1,2,3,4,5,6,7,8,9}; path: glusterfs_11.2-3/tests/basic/all_squash.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 $H0:$B0/${V0}{0,1} TEST $CLI volume set $V0 nfs.disable false path: glusterfs_11.2-3/tests/bugs/distribute/bug-1066798.t ## Start glusterd TEST glusterd; TEST pidof glusterd; TEST $CLI volume info; path: glusterfs_11.2-3/tests/features/glfs-lease.t ## Start and create a volume TEST glusterd; TEST pidof glusterd; TEST $CLI volume info; path: glusterfs_11.2-3/tests/basic/afr/granular-esh/granular-indices-but-non-granular-heal.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/${V0}{0,1} path: glusterfs_11.2-3/tests/bugs/write-behind/bug-1058663.t ## Start and create a volume TEST glusterd; TEST pidof glusterd; TEST $CLI volume info; path: glusterfs_11.2-3/tests/bugs/index/bug-1559004-EMLINK-handling.t TEST glusterd TEST pidof glusterd TEST $CLI volume info path: glusterfs_11.2-3/tests/basic/distribute/phase1-migration-fops.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 $H0:$B0/${V0}{0,1} TEST $CLI volume set $V0 performance.quick-read off path: glusterfs_11.2-3/tests/basic/changelog/changelog-history.t TEST glusterd TEST pidof glusterd sleep 3 path: glusterfs_11.2-3/tests/bugs/snapshot/bug-1178079.t #Create a volume TEST glusterd; TEST pidof glusterd; TEST $CLI volume info; path: glusterfs_11.2-3/tests/basic/afr/arbiter-remove-brick.t TEST glusterd TEST pidof glusterd #Create arbiter volume. path: glusterfs_11.2-3/tests/bitrot/bug-1700078.t ## Start glusterd TEST glusterd; TEST pidof glusterd; ## Lets create and start the volume path: glusterfs_11.2-3/tests/bugs/distribute/bug-884597.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 $H0:$B0/${V0}0 $H0:$B0/${V0}1 $H0:$B0/${V0}2 path: glusterfs_11.2-3/tests/bugs/unclassified/bug-1034085.t #Create a volume TEST glusterd; TEST pidof glusterd TEST $CLI volume info; path: glusterfs_11.2-3/tests/basic/afr/afr-no-fsync.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 3 $H0:$B0/brick{0,1,3} TEST $CLI volume set $V0 features.shard on path: glusterfs_11.2-3/tests/features/worm_sh.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/${V0}{0,1} TEST $CLI volume start $V0 path: glusterfs_11.2-3/tests/bugs/replicate/bug-1292379.t TEST glusterd TEST pidof glusterd TEST $CLI volume info path: glusterfs_11.2-3/tests/bugs/ec/bug-1708156-honor-inodelk-contention-notify-on-partial-locks.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 disperse 6 redundancy 2 $H0:$B0/${V0}{0..5} path: glusterfs_11.2-3/tests/basic/distribute/non-root-unlink-stale-linkto.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 $H0:$B0/$V0-{1,2} path: glusterfs_11.2-3/tests/bugs/replicate/bug-1417522-block-split-brain-resolution.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 3 $H0:$B0/${V0}{0..2} TEST $CLI volume set $V0 self-heal-daemon off path: glusterfs_11.2-3/tests/features/weighted-rebalance.t TEST glusterd TEST pidof glusterd TEST $CLI volume info path: glusterfs_11.2-3/tests/bugs/glusterfs-server/bug-889996.t TEST glusterd TEST pidof glusterd TEST $CLI volume info; path: glusterfs_11.2-3/tests/bugs/posix/bug-1113960.t TEST glusterd TEST pidof glusterd TEST $CLI volume info; path: glusterfs_11.2-3/tests/bugs/ec/bug-1161886.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 redundancy 1 $H0:$B0/${V0}{0..2} EXPECT "Created" volinfo_field $V0 'Status' path: glusterfs_11.2-3/tests/bugs/shard/bug-1256580.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/${V0}{0,1,2,3} TEST $CLI volume set $V0 features.shard on path: glusterfs_11.2-3/tests/bugs/core/log-bug-1362520.t #Basic checks TEST glusterd TEST pidof glusterd TEST $CLI volume info path: glusterfs_11.2-3/tests/bugs/snapshot/bug-1162498-zfs.t TEST verify_zfs_version; TEST glusterd; TEST pidof glusterd; TEST setup_zfs 1 path: glusterfs_11.2-3/tests/bugs/replicate/bug-1473026.t cleanup; TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 3 $H0:$B0/${V0}{0,1,2} path: glusterfs_11.2-3/tests/bugs/replicate/bug-1473026.t # Glusterd should start but the volume info and brick volfiles don't match TEST glusterd TEST pidof glusterd cleanup; path: glusterfs_11.2-3/tests/bugs/cli/bug-1353156-get-state-cli-validations.t TEST glusterd TEST pidof glusterd TEST mkdir -p $ODIR path: glusterfs_11.2-3/tests/basic/afr/heal-info.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/brick{0,1} TEST $CLI volume set $V0 locking-scheme granular path: glusterfs_11.2-3/tests/bugs/snapshot/bug-1227646-zfs.t TEST glusterd; TEST pidof glusterd; #TEST $CLI volume create $V0 $H0:$L1 $H0:$L2 $H0:$L3; path: glusterfs_11.2-3/tests/basic/afr/bug-1722507-type-mismatch-error-handling-non-granular.t ## Start and create a volume TEST glusterd; TEST pidof glusterd; TEST $CLI volume info; path: glusterfs_11.2-3/tests/bugs/posix/bug-990028.t { TEST glusterd TEST pidof glusterd TEST $CLI volume info; path: glusterfs_11.2-3/tests/bugs/replicate/bug-1493415-gfid-heal.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/${V0}{0,1} TEST $CLI volume start $V0 path: glusterfs_11.2-3/tests/basic/open-fd-snap-delete.t TEST glusterd; TEST pidof glusterd; TEST $CLI volume create $V0 $H0:$L1 $H0:$L2 $H0:$L3; path: glusterfs_11.2-3/tests/bugs/snapshot/bug-1279327-zfs.t TEST verify_zfs_version; TEST glusterd; TEST pidof glusterd; TEST init_n_bricks 3 path: glusterfs_11.2-3/tests/bugs/shard/bug-1248887.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/${V0}{0,1} TEST $CLI volume set $V0 features.shard on path: glusterfs_11.2-3/tests/bugs/distribute/bug-1161156.t TEST glusterd TEST pidof glusterd TEST $CLI volume info; path: glusterfs_11.2-3/tests/bugs/protocol/bug-808400.t TEST glusterd TEST pidof glusterd TEST $CLI volume info; path: glusterfs_11.2-3/tests/line-coverage/log-and-brick-ops-negative-case.t cleanup TEST glusterd TEST pidof glusterd #create volumes path: glusterfs_11.2-3/tests/bugs/replicate/bug-1728770-pass-xattrs.t TEST glusterd; TEST pidof glusterd; TEST $CLI volume create $V0 replica 3 $H0:$L1 $H0:$L2 $H0:$L3 $H0:$L4 $H0:$L5 $H0:$L6; path: glusterfs_11.2-3/tests/bugs/core/bug-908146.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 $H0:$B0/${V0}0 TEST $CLI volume set $V0 performance.open-behind off path: glusterfs_11.2-3/tests/performance/quick-read.t TEST glusterd TEST pidof glusterd TEST $CLI volume info; path: glusterfs_11.2-3/tests/bugs/access-control/bug-958691.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 $H0:$B0/${V0}{0,1} TEST $CLI volume set $V0 nfs.disable false path: glusterfs_11.2-3/tests/bugs/distribute/bug-1286171.t # Start glusterd TEST glusterd; TEST pidof glusterd; TEST $CLI volume info; path: glusterfs_11.2-3/tests/basic/afr/gfid-mismatch-resolution-with-cli.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/${V0}{0,1} TEST $CLI volume start $V0 path: glusterfs_11.2-3/tests/bugs/shard/bug-1259651.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/${V0}{0,1,2,3} TEST $CLI volume start $V0 path: glusterfs_11.2-3/tests/basic/meta.t TEST glusterd TEST pidof glusterd TEST $CLI volume info; path: glusterfs_11.2-3/tests/bugs/distribute/bug-1190734.t TEST glusterd TEST pidof glusterd # Test 1-2 Create repliacted volume path: glusterfs_11.2-3/tests/bugs/snapshot/bug-1250387.t TEST glusterd; TEST pidof glusterd; TEST $CLI volume create $V0 $H0:$L1; path: glusterfs_11.2-3/tests/bugs/core/bug-1402841.t-mt-dir-scan-race.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/${V0}{0,1} TEST $CLI volume set $V0 self-heal-daemon off path: glusterfs_11.2-3/tests/bugs/nfs/bug-1053579.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 $H0:$B0/${V0}1 TEST $CLI volume set $V0 nfs.disable false path: glusterfs_11.2-3/tests/bugs/glusterfs/bug-893378.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 $H0:$B0/${V0}0 $H0:$B0/${V0}1 $H0:$B0/${V0}2 path: glusterfs_11.2-3/tests/bugs/replicate/bug-1498570-client-iot-graph-check.t cleanup TEST glusterd TEST pidof glusterd } path: glusterfs_11.2-3/extras/command-completion/gluster.bash elif [ "$1" == "complete" ]; then if ! pidof glusterd > /dev/null 2>&1; then list=''; path: glusterfs_11.2-3/extras/command-completion/gluster.bash if pidof glusterd > /dev/null 2>&1; then GLUSTER_SET_OPTIONS=" $(for token in `gluster volume set help 2>/dev/null | grep "^Option:" | cut -d ' ' -f 2` path: glusterfs_11.2-3/tests/bugs/cli/bug-1113476.t TEST verify_lvm_version TEST glusterd TEST pidof glusterd TEST setup_lvm 1 path: glusterfs_11.2-3/tests/bugs/replicate/bug-888174.t cleanup; TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/r2_0 $H0:$B0/r2_1 path: glusterfs_11.2-3/tests/bugs/protocol/bug-762989.t ## Start and create a volume TEST glusterd TEST pidof glusterd TEST $CLI volume info; path: glusterfs_11.2-3/tests/basic/afr/split-brain-healing.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/${V0}{1,2,3,4} TEST $CLI volume set $V0 cluster.self-heal-daemon off path: glusterfs_11.2-3/tests/basic/afr/afr-read-hash-mode.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 3 arbiter 1 $H0:$B0/${V0}{0..2} path: glusterfs_11.2-3/tests/features/subdir-mount.t ## Start and create a volume TEST glusterd TEST pidof glusterd TEST $CLI volume info; path: glusterfs_11.2-3/debian/patches/05-remove-hashbang.diff -#!/bin/bash - if pidof glusterd > /dev/null 2>&1; then GLUSTER_SET_OPTIONS=" $(for token in `gluster volume set help 2>/dev/null | grep "^Option:" | cut -d ' ' -f 2` path: glusterfs_11.2-3/tests/bugs/bug-1371806_3.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 $H0:$B0/${V0}{0,1,2,3} TEST $CLI volume start $V0 path: glusterfs_11.2-3/tests/bugs/bug-1371806_2.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 $H0:$B0/${V0}{0,1,2,3} TEST $CLI volume start $V0 path: glusterfs_11.2-3/tests/bugs/bug-1371806_1.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 $H0:$B0/${V0}{0,1,2,3} TEST $CLI volume start $V0 path: glusterfs_11.2-3/tests/basic/fencing/fence-basic.t TEST glusterd TEST pidof glusterd TEST $CLI volume info; path: glusterfs_11.2-3/tests/bugs/replicate/bug-1386188-sbrain-fav-child.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/${V0}{0,1} TEST $CLI volume set $V0 self-heal-daemon off path: glusterfs_11.2-3/tests/bugs/rpc/bug-954057.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 $H0:$B0/${V0} TEST $CLI volume start $V0 path: glusterfs_11.2-3/tests/basic/fencing/afr-lock-heal-basic.t TEST glusterd TEST pidof glusterd TEST $CLI volume info; path: glusterfs_11.2-3/tests/bugs/bitrot/bug-1229134-bitd-not-support-vol-set.t ## Start glusterd TEST glusterd; TEST pidof glusterd; ## Lets create and start the volume path: glusterfs_11.2-3/tests/bugs/nfs/bug-847622.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 $H0:$B0/brick0 TEST $CLI volume set $V0 nfs.disable false path: glusterfs_11.2-3/tests/bugs/access-control/bug-887098-gmount-crash.t ## Start and create a volume TEST glusterd; TEST pidof glusterd; TEST $CLI volume info; path: glusterfs_11.2-3/tests/bugs/snapshot/bug-1250387-zfs.t TEST glusterd; TEST pidof glusterd; TEST $CLI volume create $V0 $H0:$L1; path: glusterfs_11.2-3/tests/bugs/distribute/bug-915554.t TEST glusterd TEST pidof glusterd BRICK_COUNT=3 path: glusterfs_11.2-3/tests/bugs/glusterd/bug-1085330-and-bug-916549.t kill -SIGKILL $brick_pid; TEST $CLI volume start $V0 force; TEST process_leak_count $(pidof glusterd); cleanup path: glusterfs_11.2-3/tests/bugs/glusterd/bug-1085330-and-bug-916549.t cleanup; TEST glusterd TEST pidof glusterd TEST $CLI volume info; path: glusterfs_11.2-3/tests/bugs/protocol/bug-1390914.t #test that fops are not wound on anon-fd when fd is not open on that brick TEST glusterd; TEST pidof glusterd TEST $CLI volume create $V0 replica 3 $H0:$B0/${V0}{1,2,3}; path: glusterfs_11.2-3/tests/bugs/cli/bug-1077682.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 $H0:$B0/${V0}{1,2,3,4} path: glusterfs_11.2-3/tests/bugs/io-stats/bug-1598548.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 $H0:$B0/${V0}0 TEST $CLI volume start $V0 path: glusterfs_11.2-3/tests/000-flaky/bugs_nfs_bug-1116503.t cleanup; TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 $H0:$B0/$V0 path: glusterfs_11.2-3/tests/features/unhashed-auto.t TEST glusterd TEST pidof glusterd TEST $CLI volume info path: glusterfs_11.2-3/tests/basic/op_errnos.t TEST verify_lvm_version; TEST glusterd; TEST pidof glusterd; TEST setup_lvm 1 path: glusterfs_11.2-3/tests/basic/afr/split-brain-heal-info.t SPB_FILES=0 TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/${V0}{0,1} TEST $CLI volume start $V0 path: glusterfs_11.2-3/tests/00-geo-rep/00-georep-verify-setup.t #get-state commamd shouldn't crash glusterd when geo-rep session is configured TEST $CLI get-state TEST pidof glusterd TEST $CLI get-state detail path: glusterfs_11.2-3/tests/00-geo-rep/00-georep-verify-setup.t TEST $CLI get-state detail TEST pidof glusterd #Pause geo-replication session path: glusterfs_11.2-3/tests/00-geo-rep/00-georep-verify-setup.t cleanup; TEST glusterd; TEST pidof glusterd path: glusterfs_11.2-3/tests/bugs/cli/bug-1087487.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 $H0:$B0/${V0}0 $H0:$B0/${V0}1 path: glusterfs_11.2-3/tests/basic/fencing/fencing-crash-conistency.t TEST glusterd TEST pidof glusterd TEST $CLI volume info; path: glusterfs_11.2-3/tests/basic/afr/inodelk.t TEST glusterd; TEST pidof glusterd TEST $CLI volume create $V0 replica 3 arbiter 1 $H0:$B0/${V0}{0..5} path: glusterfs_11.2-3/tests/basic/afr/inodelk.t #Do similar tests on replica 2 TEST glusterd; TEST pidof glusterd TEST $CLI volume create $V0 replica 2 $H0:$B0/${V0}{0..3} TEST $CLI volume start $V0 path: glusterfs_11.2-3/tests/basic/distribute/throttle-rebal.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 $H0:$B0/brick1 $H0:$B0/brick2 TEST $CLI volume start $V0 path: glusterfs_11.2-3/tests/bugs/replicate/bug-1637802-arbiter-stale-data-heal-lock.t TEST glusterd; TEST pidof glusterd; TEST $CLI volume create $V0 replica 3 arbiter 1 $H0:$B0/${V0}{0,1,2}; TEST $CLI volume start $V0; path: glusterfs_11.2-3/tests/bugs/distribute/bug-1247563.t #Basic checks TEST glusterd TEST pidof glusterd TEST $CLI volume info path: glusterfs_11.2-3/tests/00-geo-rep/bug-1708603.t cleanup; TEST glusterd; TEST pidof glusterd path: glusterfs_11.2-3/run-tests.sh fi # Check for pidof pidof pidof > /dev/null 2>&1 if [ $? -ne 0 ]; then path: glusterfs_11.2-3/run-tests.sh # Check for pidof pidof pidof > /dev/null 2>&1 if [ $? -ne 0 ]; then MISSING="$MISSING pidof" path: glusterfs_11.2-3/run-tests.sh pidof pidof > /dev/null 2>&1 if [ $? -ne 0 ]; then MISSING="$MISSING pidof" fi path: glusterfs_11.2-3/tests/features/fuse-lru-limit.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 $H0:$B0/${V0}{0,1} TEST $CLI volume start $V0 path: glusterfs_11.2-3/tests/basic/changelog/history-api.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 $H0:$B0/${V0}0 path: glusterfs_11.2-3/tests/bugs/replicate/bug-1657783-do-not-update-read-subvol-on-rename-link.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 replica 3 $H0:$B0/${V0}{0..2} TEST $CLI volume set $V0 self-heal-daemon off path: glusterfs_11.2-3/tests/bugs/distribute/bug-1204140.t TEST glusterd TEST pidof glusterd TEST $CLI volume create $V0 $H0:$B0/${V0}0 $H0:$B0/${V0}1 $H0:$B0/${V0}2