I don't know why I didn't think to get some more info regarding my storage
environment and post it here earlier. My gluster_vg1 volume is on /dev/sda1. I can
access the engine storage directory but I think that is because it is not thin
provisioned. I guess I was too bogged down in solving the problem when I'm stuck in
emergency mode. I had to sneaker net my USB drive to my system so I could capture some
info. Anyhow here it is:
# lsblk
NAME MAJ:MIN RM SIZE RO TYPE
MOUNTPOINT
sda 8:0 0 5.5T 0 disk
└─sda1 8:1 0 5.5T 0 part
sdb 8:16 0 223.6G 0 disk
├─sdb1 8:17 0 1G 0 part
/boot
└─sdb2 8:18 0 222.6G 0 part
└─md127 9:127 0 222.5G 0 raid1
├─onn_vmh-swap 253:0 0 4G 0 lvm
[SWAP]
├─onn_vmh-pool00_tmeta 253:1 0 1G 0 lvm
│ └─onn_vmh-pool00-tpool 253:3 0 173.6G 0 lvm
│ ├─onn_vmh-ovirt--node--ng--4.2.7.1--0.20181216.0+1 253:4 0 146.6G 0 lvm /
│ ├─onn_vmh-pool00 253:5 0 173.6G 0 lvm
│ ├─onn_vmh-root 253:6 0 146.6G 0 lvm
│ ├─onn_vmh-home 253:7 0 1G 0 lvm
/home
│ ├─onn_vmh-tmp 253:8 0 1G 0 lvm
/tmp
│ ├─onn_vmh-var 253:9 0 15G 0 lvm
/var
│ ├─onn_vmh-var_log 253:10 0 8G 0 lvm
/var/log
│ ├─onn_vmh-var_log_audit 253:11 0 2G 0 lvm
/var/log/audit
│ └─onn_vmh-var_crash 253:12 0 10G 0 lvm
/var/crash
└─onn_vmh-pool00_tdata 253:2 0 173.6G 0 lvm
└─onn_vmh-pool00-tpool 253:3 0 173.6G 0 lvm
├─onn_vmh-ovirt--node--ng--4.2.7.1--0.20181216.0+1 253:4 0 146.6G 0 lvm /
├─onn_vmh-pool00 253:5 0 173.6G 0 lvm
├─onn_vmh-root 253:6 0 146.6G 0 lvm
├─onn_vmh-home 253:7 0 1G 0 lvm
/home
├─onn_vmh-tmp 253:8 0 1G 0 lvm
/tmp
├─onn_vmh-var 253:9 0 15G 0 lvm
/var
├─onn_vmh-var_log 253:10 0 8G 0 lvm
/var/log
├─onn_vmh-var_log_audit 253:11 0 2G 0 lvm
/var/log/audit
└─onn_vmh-var_crash 253:12 0 10G 0 lvm
/var/crash
sdc 8:32 0 223.6G 0 disk
└─sdc1 8:33 0 222.6G 0 part
└─md127 9:127 0 222.5G 0 raid1
├─onn_vmh-swap 253:0 0 4G 0 lvm
[SWAP]
├─onn_vmh-pool00_tmeta 253:1 0 1G 0 lvm
│ └─onn_vmh-pool00-tpool 253:3 0 173.6G 0 lvm
│ ├─onn_vmh-ovirt--node--ng--4.2.7.1--0.20181216.0+1 253:4 0 146.6G 0 lvm /
│ ├─onn_vmh-pool00 253:5 0 173.6G 0 lvm
│ ├─onn_vmh-root 253:6 0 146.6G 0 lvm
│ ├─onn_vmh-home 253:7 0 1G 0 lvm
/home
│ ├─onn_vmh-tmp 253:8 0 1G 0 lvm
/tmp
│ ├─onn_vmh-var 253:9 0 15G 0 lvm
/var
│ ├─onn_vmh-var_log 253:10 0 8G 0 lvm
/var/log
│ ├─onn_vmh-var_log_audit 253:11 0 2G 0 lvm
/var/log/audit
│ └─onn_vmh-var_crash 253:12 0 10G 0 lvm
/var/crash
└─onn_vmh-pool00_tdata 253:2 0 173.6G 0 lvm
└─onn_vmh-pool00-tpool 253:3 0 173.6G 0 lvm
├─onn_vmh-ovirt--node--ng--4.2.7.1--0.20181216.0+1 253:4 0 146.6G 0 lvm /
├─onn_vmh-pool00 253:5 0 173.6G 0 lvm
├─onn_vmh-root 253:6 0 146.6G 0 lvm
├─onn_vmh-home 253:7 0 1G 0 lvm
/home
├─onn_vmh-tmp 253:8 0 1G 0 lvm
/tmp
├─onn_vmh-var 253:9 0 15G 0 lvm
/var
├─onn_vmh-var_log 253:10 0 8G 0 lvm
/var/log
├─onn_vmh-var_log_audit 253:11 0 2G 0 lvm
/var/log/audit
└─onn_vmh-var_crash 253:12 0 10G 0 lvm
/var/crash
sdd 8:48 0 596.2G 0 disk
└─sdd1 8:49 0 4G 0 part
└─gluster_vg3-tmpLV 253:13 0 2G 0 lvm
sde 8:64 1 7.5G 0 disk
└─sde1 8:65 1 7.5G 0 part
/mnt
# blkid
/dev/sda1: UUID="f026a2dc-201a-4b43-974e-2419a8783bce" TYPE="xfs"
PARTLABEL="Linux filesystem"
PARTUUID="4bca8a3a-42f0-4877-aa60-f544bf1fdce7"
/dev/sdc1: UUID="e5f4acf5-a4bc-6470-7b6f-415e3f4077ff"
UUID_SUB="a895900e-5585-8f31-7515-1ff7534e39d7"
LABEL="vmh.cyber-range.lan:pv00" TYPE="linux_raid_member"
/dev/sdb1: UUID="9b9546f9-25d2-42a6-835b-303f32aee4b1" TYPE="ext4"
/dev/sdb2: UUID="e5f4acf5-a4bc-6470-7b6f-415e3f4077ff"
UUID_SUB="6e20b5dd-0152-7f42-22a7-c17133fbce45"
LABEL="vmh.cyber-range.lan:pv00" TYPE="linux_raid_member"
/dev/sdd1: UUID="2nLjVF-sh3N-0qkm-aUQ1-jnls-3e8W-tUkBw5"
TYPE="LVM2_member"
/dev/md127: UUID="Mq1chn-6XhF-WCwF-LYhl-tZEz-Y8lq-8R2Ifq"
TYPE="LVM2_member"
/dev/mapper/onn_vmh-swap: UUID="1b0b9c91-22ed-41d1-aebf-e22fd9aa05d9"
TYPE="swap"
/dev/mapper/onn_vmh-ovirt--node--ng--4.2.7.1--0.20181216.0+1:
UUID="b0e1c479-9696-4e19-b799-7f81236026b7" TYPE="ext4"
/dev/mapper/onn_vmh-root: UUID="60905f5d-ed91-4ca9-9729-9a72a4678ddd"
TYPE="ext4"
/dev/mapper/onn_vmh-home: UUID="82a1d567-f8af-4b96-bfbf-5f79dff7384f"
TYPE="ext4"
/dev/mapper/onn_vmh-tmp: UUID="7dd9d3ae-3af7-4763-9683-19f583d8d15b"
TYPE="ext4"
/dev/mapper/onn_vmh-var: UUID="f206e030-876b-45a9-8a90-a0e54005b85c"
TYPE="ext4"
/dev/mapper/onn_vmh-var_log: UUID="b8a12f56-0818-416c-9fb7-33b48ef29eed"
TYPE="ext4"
/dev/mapper/onn_vmh-var_log_audit: UUID="bc78ad0c-9ab6-4f57-a69f-5b1ddf898552"
TYPE="ext4"
/dev/mapper/onn_vmh-var_crash: UUID="a941d416-4d7d-41ae-bcd4-8c1ec9d0f744"
TYPE="ext4"
/dev/sde1: UUID="44aa40d0-6c82-4e8e-8218-177e5c8474f4" TYPE="ext4"
# pvscan
PV /dev/md127 VG onn_vmh lvm2 [222.44 GiB / 43.66 GiB free]
PV /dev/sdd1 VG gluster_vg3 lvm2 [<4.00 GiB / <2.00 GiB free]
Total: 2 [<226.44 GiB] / in use: 2 [<226.44 GiB] / in no VG: 0 [0 ]
Reading all physical volumes. This may take a while...
Found volume group "onn_vmh" using metadata type lvm2
Found volume group "gluster_vg3" using metadata type lvm2
# lvscan
ACTIVE '/dev/onn_vmh/pool00' [173.60 GiB] inherit
ACTIVE '/dev/onn_vmh/root' [146.60 GiB] inherit
ACTIVE '/dev/onn_vmh/home' [1.00 GiB] inherit
ACTIVE '/dev/onn_vmh/tmp' [1.00 GiB] inherit
ACTIVE '/dev/onn_vmh/var' [15.00 GiB] inherit
ACTIVE '/dev/onn_vmh/var_log' [8.00 GiB] inherit
ACTIVE '/dev/onn_vmh/var_log_audit' [2.00 GiB] inherit
ACTIVE '/dev/onn_vmh/swap' [4.00 GiB] inherit
inactive '/dev/onn_vmh/ovirt-node-ng-4.2.7.1-0.20181216.0' [146.60 GiB]
inherit
ACTIVE '/dev/onn_vmh/ovirt-node-ng-4.2.7.1-0.20181216.0+1' [146.60
GiB] inherit
ACTIVE '/dev/onn_vmh/var_crash' [10.00 GiB] inherit
ACTIVE '/dev/gluster_vg3/tmpLV' [2.00 GiB] inherit
[/etc/lvm/backup/gluster_vg1]
# Generated by LVM2 version 2.02.180(2)-RHEL7 (2018-07-20): Sat Dec 22 10:18:46 2018
contents = "Text Format Volume Group"
version = 1
description = "Created *after* executing '/usr/sbin/lvcreate --virtualsize 500GB
--name lv_datadisks -T gluster_vg1/lvthinpool'"
creation_host = "vmh.cyber-range.lan" # Linux vmh.cyber-range.lan
3.10.0-957.1.3.el7.x86_64 #1 SMP Thu Nov 29 14:49:43 UTC 2018 x86_64
creation_time = 1545495526 # Sat Dec 22 10:18:46 2018
gluster_vg1 {
id = "TfNqtn-2eX6-i5gC-w4ye-h29n-5Zfy-UFHSvU"
seqno = 9
format = "lvm2" # informational
status = ["RESIZEABLE", "READ", "WRITE"]
flags = []
extent_size = 2048 # 1024 Kilobytes
max_lv = 0
max_pv = 0
metadata_copies = 0
physical_volumes {
pv0 {
id = "DRfoKl-TUhb-cirx-Oz9P-mZEY-XoiB-mdLy6v"
device = "/dev/sda" # Hint only
status = ["ALLOCATABLE"]
flags = []
dev_size = 11718885376 # 5.45703 Terabytes
pe_start = 2048
pe_count = 5722111 # 5.45703 Terabytes
}
}
logical_volumes {
engine_lv {
id = "P2ScEB-ws3V-iqVv-XTWh-Y2Jh-pg0c-6KnJXC"
status = ["READ", "WRITE", "VISIBLE"]
flags = []
creation_time = 1545495460 # 2018-12-22 10:17:40 -0600
creation_host = "vmh.cyber-range.lan"
segment_count = 1
segment1 {
start_extent = 0
extent_count = 102400 # 100 Gigabytes
type = "striped"
stripe_count = 1 # linear
stripes = [
"pv0", 0
]
}
}
lvthinpool {
id = "c0yaNn-DcaB-cYjj-9ZRv-M2Em-rzLW-qY9WpI"
status = ["READ", "WRITE", "VISIBLE"]
flags = []
creation_time = 1545495487 # 2018-12-22 10:18:07 -0600
creation_host = "vmh.cyber-range.lan"
segment_count = 1
segment1 {
start_extent = 0
extent_count = 512000 # 500 Gigabytes
type = "thin-pool"
metadata = "lvthinpool_tmeta"
pool = "lvthinpool_tdata"
transaction_id = 2
chunk_size = 2048 # 1024 Kilobytes
discards = "passdown"
zero_new_blocks = 1
}
}
lv_vmdisks {
id = "erpXRi-nPUq-mCf2-ga2J-3a0l-OWiC-i0xr8M"
status = ["READ", "WRITE", "VISIBLE"]
flags = []
creation_time = 1545495493 # 2018-12-22 10:18:13 -0600
creation_host = "vmh.cyber-range.lan"
segment_count = 1
segment1 {
start_extent = 0
extent_count = 4613735 # 4.4 Terabytes
type = "thin"
thin_pool = "lvthinpool"
transaction_id = 0
device_id = 1
}
}
lv_datadisks {
id = "hKim3z-1QCh-dwhU-st2O-t4tG-wIss-UpLMZw"
status = ["READ", "WRITE", "VISIBLE"]
flags = []
creation_time = 1545495526 # 2018-12-22 10:18:46 -0600
creation_host = "vmh.cyber-range.lan"
segment_count = 1
segment1 {
start_extent = 0
extent_count = 512000 # 500 Gigabytes
type = "thin"
thin_pool = "lvthinpool"
transaction_id = 1
device_id = 2
}
}
lvol0_pmspare {
id = "bHc0eC-Z4Ed-mV47-QTU4-SCvo-FWbE-L8NV7Q"
status = ["READ", "WRITE"]
flags = []
creation_time = 1545495487 # 2018-12-22 10:18:07 -0600
creation_host = "vmh.cyber-range.lan"
segment_count = 1
segment1 {
start_extent = 0
extent_count = 16192 # 15.8125 Gigabytes
type = "striped"
stripe_count = 1 # linear
stripes = [
"pv0", 102400
]
}
}
lvthinpool_tmeta {
id = "WBut10-rAOP-FzA7-bJvr-ZdxL-lB70-jzz1Tv"
status = ["READ", "WRITE"]
flags = []
creation_time = 1545495487 # 2018-12-22 10:18:07 -0600
creation_host = "vmh.cyber-range.lan"
segment_count = 1
segment1 {
start_extent = 0
extent_count = 16192 # 15.8125 Gigabytes
type = "striped"
stripe_count = 1 # linear
stripes = [
"pv0", 630592
]
}
}
lvthinpool_tdata {
id = "rwNZux-1fz1-dv8J-yN2j-LcES-f6ml-231td5"
status = ["READ", "WRITE"]
flags = []
creation_time = 1545495487 # 2018-12-22 10:18:07 -0600
creation_host = "vmh.cyber-range.lan"
segment_count = 1
segment1 {
start_extent = 0
extent_count = 512000 # 500 Gigabytes
type = "striped"
stripe_count = 1 # linear
stripes = [
"pv0", 118592
]
}
}
}
}
# cd /var/log
# grep -ri gluster_vg1-lvthinpool-tpool
messages-20190922:Sep 15 03:58:15 vmh lvm[14072]: Failed command for
gluster_vg1-lvthinpool-tpool.
messages:Sep 22 03:44:05 vmh lvm[14072]: Failed command for gluster_vg1-lvthinpool-tpool.
messages-20190908:Sep 1 21:27:14 vmh lvm[14062]: Monitoring thin pool
gluster_vg1-lvthinpool-tpool.
messages-20190908:Sep 1 21:27:24 vmh lvm[14062]: WARNING: Thin pool
gluster_vg1-lvthinpool-tpool data is now 100.00% full.
messages-20190908:Sep 2 00:19:05 vmh lvm[14072]: Monitoring thin pool
gluster_vg1-lvthinpool-tpool.
messages-20190908:Sep 2 00:19:15 vmh lvm[14072]: WARNING: Thin pool
gluster_vg1-lvthinpool-tpool data is now 100.00% full.
messages-20190908:Sep 2 20:16:34 vmh lvm[14072]: Failed command for
gluster_vg1-lvthinpool-tpool.