e789619b34
This patch brings some of the physical and virtual drive attributes as `custom_fields` to the disks inventory. The goal is to have this information present to ease disks maintenance when a drive becomes unavailable and its attributes can't be read anymore from the RAID controller. It also helps to standardize the extended disk attributes across the different manufacturers. As the disk physical identifers were not available under the correct format (hexadecimal format using the `xml` output as opposed as `X:Y:Z` format using the default `list` format), the command line parser has been refactored to read the `list` format, rather than `xml` one in the `omreport` raid controller parser. As the custom fields have to be created prior being able to register the disks extended attributes, this feature is only activated using the `--process-virtual-drives` command line parameter, or by setting `process_virtual_drives` to `true` in the configuration file. The custom fields to create as `DCIM > inventory item` `Text` are described below. NAME LABEL DESCRIPTION mount_point Mount point Device mount point(s) pd_identifier Physical disk identifier Physical disk identifier in the RAID controller vd_array Virtual drive array Virtual drive array the disk is member of vd_consistency Virtual drive consistency Virtual disk array consistency vd_device Virtual drive device Virtual drive system device vd_raid_type Virtual drive RAID Virtual drive array RAID type vd_size Virtual drive size Virtual drive array size In the current implementation, the disks attributes ore not updated: if a disk with the correct serial number is found, it's sufficient to consider it as up to date. To force the reprocessing of the disks extended attributes, the `--force-disk-refresh` command line option can be used: it removes all existing disks to before populating them with the correct parsing. Unless this option is specified, the extended attributes won't be modified unless a disk is replaced. It is possible to dump the physical/virtual disks map on the filesystem under the JSON notation to ease or automate disks management. The file path has to be provided using the `--dump-disks-map` command line parameter.
181 lines
5.9 KiB
Python
181 lines
5.9 KiB
Python
from netbox_agent.misc import is_tool
|
|
import subprocess
|
|
import logging
|
|
import json
|
|
import sys
|
|
|
|
|
|
class LSHW():
|
|
def __init__(self):
|
|
if not is_tool('lshw'):
|
|
logging.error('lshw does not seem to be installed')
|
|
sys.exit(1)
|
|
|
|
data = subprocess.getoutput(
|
|
'lshw -quiet -json'
|
|
)
|
|
json_data = json.loads(data)
|
|
# Starting from version 02.18, `lshw -json` wraps its result in a list
|
|
# rather than returning directly a dictionary
|
|
if isinstance(json_data, list):
|
|
self.hw_info = json_data[0]
|
|
else:
|
|
self.hw_info = json_data
|
|
self.info = {}
|
|
self.memories = []
|
|
self.interfaces = []
|
|
self.cpus = []
|
|
self.power = []
|
|
self.disks = []
|
|
self.gpus = []
|
|
self.vendor = self.hw_info["vendor"]
|
|
self.product = self.hw_info["product"]
|
|
self.chassis_serial = self.hw_info["serial"]
|
|
self.motherboard_serial = self.hw_info["children"][0].get("serial", "No S/N")
|
|
self.motherboard = self.hw_info["children"][0].get("product", "Motherboard")
|
|
|
|
for k in self.hw_info["children"]:
|
|
if k["class"] == "power":
|
|
# self.power[k["id"]] = k
|
|
self.power.append(k)
|
|
|
|
if "children" in k:
|
|
for j in k["children"]:
|
|
if j["class"] == "generic":
|
|
continue
|
|
|
|
if j["class"] == "storage":
|
|
self.find_storage(j)
|
|
|
|
if j["class"] == "memory":
|
|
self.find_memories(j)
|
|
|
|
if j["class"] == "processor":
|
|
self.find_cpus(j)
|
|
|
|
if j["class"] == "bridge":
|
|
self.walk_bridge(j)
|
|
|
|
def get_hw_linux(self, hwclass):
|
|
if hwclass == "cpu":
|
|
return self.cpus
|
|
if hwclass == "gpu":
|
|
return self.gpus
|
|
if hwclass == "network":
|
|
return self.interfaces
|
|
if hwclass == 'storage':
|
|
return self.disks
|
|
if hwclass == 'memory':
|
|
return self.memories
|
|
|
|
def find_network(self, obj):
|
|
d = {}
|
|
d["name"] = obj["logicalname"]
|
|
d["macaddress"] = obj["serial"]
|
|
d["serial"] = obj["serial"]
|
|
d["product"] = obj["product"]
|
|
d["vendor"] = obj["vendor"]
|
|
d["description"] = obj["description"]
|
|
|
|
self.interfaces.append(d)
|
|
|
|
def find_storage(self, obj):
|
|
if "children" in obj:
|
|
for device in obj["children"]:
|
|
self.disks.append({
|
|
"logicalname": device.get("logicalname"),
|
|
"product": device.get("product"),
|
|
"serial": device.get("serial"),
|
|
"version": device.get("version"),
|
|
"size": device.get("size"),
|
|
"description": device.get("description"),
|
|
"type": device.get("description"),
|
|
})
|
|
elif "nvme" in obj["configuration"]["driver"]:
|
|
if not is_tool('nvme'):
|
|
logging.error('nvme-cli >= 1.0 does not seem to be installed')
|
|
return
|
|
try:
|
|
nvme = json.loads(
|
|
subprocess.check_output(
|
|
["nvme", '-list', '-o', 'json'],
|
|
encoding='utf8')
|
|
)
|
|
for device in nvme["Devices"]:
|
|
d = {
|
|
'logicalname': device["DevicePath"],
|
|
'product': device["ModelNumber"],
|
|
'serial': device["SerialNumber"],
|
|
"version": device["Firmware"],
|
|
'description': "NVME",
|
|
'type': "NVME",
|
|
}
|
|
if "UsedSize" in device:
|
|
d['size'] = device["UsedSize"]
|
|
if "UsedBytes" in device:
|
|
d['size'] = device["UsedBytes"]
|
|
self.disks.append(d)
|
|
except Exception:
|
|
pass
|
|
|
|
def find_cpus(self, obj):
|
|
if "product" in obj:
|
|
c = {}
|
|
c["product"] = obj["product"]
|
|
c["vendor"] = obj["vendor"]
|
|
c["description"] = obj["description"]
|
|
c["location"] = obj["slot"]
|
|
|
|
self.cpus.append(c)
|
|
|
|
def find_memories(self, obj):
|
|
if "children" not in obj:
|
|
# print("not a DIMM memory.")
|
|
return
|
|
|
|
for dimm in obj["children"]:
|
|
if "empty" in dimm["description"]:
|
|
continue
|
|
|
|
d = {}
|
|
d["slot"] = dimm.get("slot")
|
|
d["description"] = dimm.get("description")
|
|
d["id"] = dimm.get("id")
|
|
d["serial"] = dimm.get("serial", 'N/A')
|
|
d["vendor"] = dimm.get("vendor", 'N/A')
|
|
d["product"] = dimm.get("product", 'N/A')
|
|
d["size"] = dimm.get("size", 0) / 2 ** 20 / 1024
|
|
|
|
self.memories.append(d)
|
|
|
|
def find_gpus(self, obj):
|
|
if "product" in obj:
|
|
c = {}
|
|
c["product"] = obj["product"]
|
|
c["vendor"] = obj["vendor"]
|
|
c["description"] = obj["description"]
|
|
|
|
self.gpus.append(c)
|
|
|
|
def walk_bridge(self, obj):
|
|
if "children" not in obj:
|
|
return
|
|
|
|
for bus in obj["children"]:
|
|
if bus["class"] == "storage":
|
|
self.find_storage(bus)
|
|
if bus["class"] == "display":
|
|
self.find_gpus(bus)
|
|
|
|
if "children" in bus:
|
|
for b in bus["children"]:
|
|
if b["class"] == "storage":
|
|
self.find_storage(b)
|
|
if b["class"] == "network":
|
|
self.find_network(b)
|
|
if b["class"] == "display":
|
|
self.find_gpus(b)
|
|
|
|
|
|
if __name__ == "__main__":
|
|
pass
|