mirror of
https://github.com/projectacrn/acrn-hypervisor.git
synced 2025-06-02 12:25:31 +00:00
With a bunch of dramatic improvement to the schema of scenario XML recently, it is now non-trivial for users of previous releases to migrate their scenario definitions to the next-gen schema, which will be a major obstacle when upgrading to the next ACRN release. In order to ease the upgrade of scenario XML, this patch introduces a script that takes a scenario XML of previous releases and generates a new one which satisfies the latest schema. The major data that can be migrated include: - Virtual UART and shared memory configurations - VM types, load_order and guest flags Other data are preserved as long as they are still needed according to the schema. The script to print a list that summarizes the data that are unintendedly discarded during the upgrade, so that users can double check the results and edit the generated scenario using the configurator. The upgrader share the same command-line interface of the default value populator and scenario validator. More upgrading logic will be added at the same time the schema is changed. v1 -> v2: * The upgrader now pretty-prints the upgraded XML. * Apply the upgrader to the current scenario XMLs in the repo. The ordering of nodes are adjusted to align with the schema, but no essential changes are made. * Set the log level to INFO. Tracked-On: #6690 Signed-off-by: Junjie Mao <junjie.mao@intel.com>
509 lines
22 KiB
Python
Executable File
509 lines
22 KiB
Python
Executable File
#!/usr/bin/env python3
|
|
#
|
|
# Copyright (C) 2022 Intel Corporation.
|
|
#
|
|
# SPDX-License-Identifier: BSD-3-Clause
|
|
#
|
|
|
|
import os
|
|
import argparse, logging
|
|
import re
|
|
from functools import lru_cache, partialmethod
|
|
from collections import defaultdict, namedtuple
|
|
import lxml.etree as etree
|
|
|
|
from scenario_transformer import ScenarioTransformer
|
|
|
|
from pipeline import PipelineObject, PipelineStage, PipelineEngine
|
|
from lxml_loader import LXMLLoadStage
|
|
from schema_slicer import SlicingSchemaByVMTypeStage
|
|
|
|
class VirtualUartConnections:
|
|
class VirtualUartEndpoint:
|
|
# The BDF of PCI virtual UARTs starts from 00:10.0
|
|
next_dev = defaultdict(lambda: 16)
|
|
|
|
@classmethod
|
|
def from_endpoint_definition(cls, element):
|
|
# For v2.x style scenario XML, the name of the VM is a child of a `vm` node.
|
|
vm_name = element.xpath("ancestor::vm/name/text()").pop()
|
|
if "legacy" in element.tag:
|
|
base = element.find("base").text
|
|
io_port = \
|
|
"0x3F8" if base.endswith("COM1_BASE") else \
|
|
"0x2F8" if base.endswith("COM2_BASE") else \
|
|
"0x3E8" if base.endswith("COM3_BASE") else \
|
|
"0x2E8" if base.endswith("COM4_BASE") else \
|
|
base
|
|
return cls(vm_name, io_port = io_port)
|
|
else:
|
|
dev = cls.next_dev[vm_name]
|
|
cls.next_dev[vm_name] += 1
|
|
return cls(vm_name, pci_bdf = f"00:{dev:02x}.0")
|
|
|
|
def __init__(self, vm_name, io_port = None, pci_bdf = None):
|
|
self.vm_name = vm_name
|
|
self.io_port = io_port
|
|
self.pci_bdf = pci_bdf
|
|
|
|
class VirtualUartConnection:
|
|
next_id = 1
|
|
|
|
@classmethod
|
|
def from_connection_definition(cls, element):
|
|
name = element.find("name").text
|
|
ty = element.find("type").text
|
|
conn = cls(name = name, ty = ty)
|
|
for endpoint in element.findall("endpoint"):
|
|
vm_name_node = endpoint.find("vm_name")
|
|
vm_name = vm_name_node.text if vm_name_node is not None else ""
|
|
io_port_node = endpoint.find("io_port")
|
|
io_port = io_port_node.text if io_port_node is not None else None
|
|
vbdf_node = endpoint.find("vbdf")
|
|
vbdf = vbdf_node.text if vbdf_node is not None else None
|
|
conn.add_endpoint(VirtualUartConnections.VirtualUartEndpoint(vm_name, io_port, vbdf))
|
|
return conn
|
|
|
|
def __init__(self, name = None, ty = "legacy"):
|
|
if name:
|
|
self.name = name
|
|
else:
|
|
self.name = f"vUART connection {self.next_id}"
|
|
self.__class__.next_id += 1
|
|
self.ty = ty
|
|
self.endpoints = []
|
|
|
|
def add_endpoint(self, endpoint):
|
|
self.endpoints.append(endpoint)
|
|
|
|
def __init__(self):
|
|
self.conns = [] # List of connections
|
|
self.dangling_conns = {} # (vm_id, vuart_id) -> conn whose target is the key
|
|
|
|
def add_endpoint(self, element):
|
|
"""Parse the vUART endpoint definition in ACRN v2.x. Returns True if and only if the element is parsed properly."""
|
|
|
|
try:
|
|
key = (element.xpath("ancestor::vm/@id").pop(), element.xpath("@id").pop())
|
|
if key in self.dangling_conns.keys():
|
|
conn = self.dangling_conns.pop(key)
|
|
conn.add_endpoint(self.VirtualUartEndpoint.from_endpoint_definition(element))
|
|
self.conns.append(conn)
|
|
else:
|
|
ty = "legacy" if "legacy" in element.tag else "pci"
|
|
conn = self.VirtualUartConnection(ty = ty)
|
|
conn.add_endpoint(self.VirtualUartEndpoint.from_endpoint_definition(element))
|
|
self.dangling_conns[(element.xpath("target_vm_id/text()").pop(), element.xpath("target_uart_id/text()").pop())] = conn
|
|
|
|
return True
|
|
except Exception as e:
|
|
# Skip vUART endpoint definition not satisfying the schema. The discarded-data warnings will report those
|
|
# unmigrated data.
|
|
logging.debug(e)
|
|
return False
|
|
|
|
def add_connection(self, element):
|
|
"""Parse the vUART connection definition in ACRN v3.x"""
|
|
self.conns.append(self.VirtualUartConnection.from_connection_definition(element))
|
|
return True
|
|
|
|
def format_xml_elements(self, xsd_element_node):
|
|
new_parent_node = etree.Element(xsd_element_node.get("name"))
|
|
for conn in self.conns:
|
|
new_node = etree.Element("vuart_connection")
|
|
etree.SubElement(new_node, "name").text = conn.name
|
|
etree.SubElement(new_node, "type").text = conn.ty
|
|
for endpoint in conn.endpoints:
|
|
new_endpoint_node = etree.SubElement(new_node, "endpoint")
|
|
etree.SubElement(new_endpoint_node, "vm_name").text = endpoint.vm_name
|
|
if endpoint.io_port:
|
|
etree.SubElement(new_endpoint_node, "io_port").text = endpoint.io_port
|
|
if endpoint.pci_bdf:
|
|
etree.SubElement(new_endpoint_node, "vbdf").text = endpoint.pci_bdf
|
|
new_parent_node.append(new_node)
|
|
return [new_parent_node]
|
|
|
|
class SharedMemoryRegions:
|
|
class SharedMemoryRegion(namedtuple("SharedMemoryRegion", ["name", "size", "shared_vms"])):
|
|
# The BDF of IVSHMEM PCI functions starts from 00:08.0
|
|
next_dev = defaultdict(lambda: 8)
|
|
|
|
@classmethod
|
|
def from_hypervisor_encoding(cls, text, old_xml_etree):
|
|
parts = [p.strip() for p in text[text.find("/") + 1 :].split(",")]
|
|
name = parts[0]
|
|
size = parts[1]
|
|
shared_vm_ids = parts[2].split(":")
|
|
|
|
shared_vms = []
|
|
for vm_id in shared_vm_ids:
|
|
vm_name_node = old_xml_etree.xpath(f"//vm[@id='{vm_id}']/name")
|
|
if not vm_name_node:
|
|
logging.warning(f"VM {vm_id}, which is referred by shared memory region {name}, is not defined in the scenario.")
|
|
continue
|
|
|
|
vm_name = vm_name_node[0].text
|
|
dev = cls.next_dev[vm_name]
|
|
cls.next_dev[vm_name] += 1
|
|
shared_vms.append((vm_name, f"00:{dev:02x}.0"))
|
|
|
|
return cls(name, size, shared_vms)
|
|
|
|
@classmethod
|
|
def from_xml_node(cls, node):
|
|
name = node.get("name")
|
|
size = node.find("IVSHMEM_SIZE").text
|
|
shared_vms = []
|
|
for shared_vm_node in node.find("IVSHMEM_VMS"):
|
|
vm_name = shared_vm_node.find("VM_NAME").text
|
|
vbdf = shared_vm_node.find("VBDF").text
|
|
shared_vms.append((vm_name, vbdf))
|
|
return cls(name, size, shared_vms)
|
|
|
|
def format_xml_element(self):
|
|
node = etree.Element("IVSHMEM_REGION")
|
|
node.set("name", self.name)
|
|
etree.SubElement(node, "IVSHMEM_SIZE").text = self.size
|
|
|
|
vms_node = etree.SubElement(node, "IVSHMEM_VMS")
|
|
for vm_name, vbdf in self.shared_vms:
|
|
vm_node = etree.SubElement(vms_node, "IVSHMEM_VM")
|
|
etree.SubElement(vm_node, "VM_NAME").text = vm_name
|
|
etree.SubElement(vm_node, "VBDF").text = vbdf
|
|
|
|
return node
|
|
|
|
def __init__(self, old_xml_etree):
|
|
self.old_xml_etree = old_xml_etree
|
|
self.regions = {}
|
|
|
|
def add_ivshmem_region(self, ivshmem_region_node):
|
|
"""Parse IVSHMEM_REGION nodes in either v2.x and v3.x format."""
|
|
|
|
if len(ivshmem_region_node) == 0:
|
|
# ACRN v2.x format
|
|
region = self.SharedMemoryRegion.from_hypervisor_encoding(ivshmem_region_node.text, self.old_xml_etree)
|
|
self.regions[region.name] = region
|
|
else:
|
|
# ACRN v3.x format
|
|
region = self.SharedMemoryRegion.from_xml_node(ivshmem_region_node)
|
|
self.regions[region.name] = region
|
|
|
|
def format_xml_element(self):
|
|
node = etree.Element("IVSHMEM")
|
|
for region in self.regions.values():
|
|
node.append(region.format_xml_element())
|
|
return node
|
|
|
|
class ScenarioUpgrader(ScenarioTransformer):
|
|
@classmethod
|
|
def get_node(cls, element, xpath):
|
|
return next(iter(element.xpath(xpath, namespaces=cls.xpath_ns)), None)
|
|
|
|
def __init__(self, xsd_etree, old_xml_etree):
|
|
super().__init__(xsd_etree, visit_optional_node=True)
|
|
self.old_xml_etree = old_xml_etree
|
|
|
|
# Collect all nodes in old_xml_etree which will be used to track data not moved
|
|
self.old_data_nodes = set()
|
|
for node in old_xml_etree.iter():
|
|
if node.text:
|
|
self.old_data_nodes.add(node)
|
|
|
|
self.hv_vm_node_map = {}
|
|
|
|
def get_from_old_data(self, new_parent_node, xpath):
|
|
hv_vm_node = new_parent_node
|
|
if hv_vm_node.tag not in ["vm", "hv"]:
|
|
hv_vm_node = next(new_parent_node.iterancestors(["vm", "hv"]), None)
|
|
old_hv_vm_node = self.hv_vm_node_map[hv_vm_node]
|
|
old_data_node = old_hv_vm_node.xpath(xpath)
|
|
return old_data_node
|
|
|
|
def move_legacy_vuart(self, xsd_element_node, xml_parent_node, new_nodes):
|
|
# Preserve the legacy vuart for console only.
|
|
legacy_vuart = self.get_from_old_data(xml_parent_node, ".//legacy_vuart[@id = '0']")
|
|
if legacy_vuart:
|
|
new_nodes.append(legacy_vuart[0])
|
|
for child in legacy_vuart[0].iter():
|
|
self.old_data_nodes.discard(child)
|
|
return False
|
|
|
|
def move_vuart_connections(self, xsd_element_node, xml_parent_node, new_nodes):
|
|
conns = VirtualUartConnections()
|
|
|
|
# Fetch vUART endpoints in the old data
|
|
vuart_endpoints = self.old_xml_etree.xpath("//legacy_vuart[@id != '0' and base != 'INVALID_COM_BASE'] | //communication_vuart[base != 'INVALID_PCI_BASE']")
|
|
vuart_connections = self.old_xml_etree.xpath("//vuart_connection")
|
|
|
|
for endpoint in vuart_endpoints:
|
|
if conns.add_endpoint(endpoint):
|
|
for child in endpoint.iter():
|
|
self.old_data_nodes.discard(child)
|
|
|
|
for connection in vuart_connections:
|
|
if conns.add_connection(connection):
|
|
for child in connection.iter():
|
|
self.old_data_nodes.discard(child)
|
|
|
|
new_nodes.extend(conns.format_xml_elements(xsd_element_node))
|
|
|
|
# Disconnected endpoints do not migrate, but remove such nodes from old_data_nodes to avoid raising
|
|
# data-is-discarded warnings.
|
|
for n in self.old_xml_etree.xpath("//legacy_vuart[@id != '0' and base = 'INVALID_COM_BASE'] | //communication_vuart[base = 'INVALID_PCI_BASE']"):
|
|
for child in n.iter():
|
|
self.old_data_nodes.discard(child)
|
|
|
|
return False
|
|
|
|
def move_ivshmem(self, xsd_element_node, xml_parent_node, new_nodes):
|
|
regions = SharedMemoryRegions(self.old_xml_etree)
|
|
for old_region in self.old_xml_etree.xpath("//IVSHMEM_REGION"):
|
|
regions.add_ivshmem_region(old_region)
|
|
for child in old_region.iter():
|
|
self.old_data_nodes.discard(child)
|
|
new_nodes.append(regions.format_xml_element())
|
|
|
|
return False
|
|
|
|
def move_vm_type(self, xsd_element_node, xml_parent_node, new_nodes):
|
|
old_vm_type_node = self.get_from_old_data(xml_parent_node, ".//vm_type").pop()
|
|
old_guest_flag_nodes = self.get_from_old_data(xml_parent_node, ".//guest_flag[text() = 'GUEST_FLAG_RT']")
|
|
|
|
new_node = etree.Element(xsd_element_node.get("name"))
|
|
if old_vm_type_node.text in ["PRE_RT_VM", "POST_RT_VM"] or old_guest_flag_nodes:
|
|
new_node.text = "RTVM"
|
|
elif old_vm_type_node.text in ["SAFETY_VM", "PRE_STD_VM", "POST_STD_VM"]:
|
|
new_node.text = "STANDARD_VM"
|
|
else:
|
|
new_node.text = old_vm_type_node.text
|
|
new_nodes.append(new_node)
|
|
self.old_data_nodes.discard(old_vm_type_node)
|
|
for n in old_guest_flag_nodes:
|
|
self.old_data_nodes.discard(n)
|
|
|
|
return False
|
|
|
|
def move_guest_flag(self, guest_flag, xsd_element_node, xml_parent_node, new_nodes):
|
|
old_data_nodes = self.get_from_old_data(xml_parent_node, f".//guest_flag[text() = '{guest_flag}']")
|
|
if old_data_nodes:
|
|
new_node = etree.Element(xsd_element_node.get("name"))
|
|
new_node.text = "y"
|
|
new_nodes.append(new_node)
|
|
for n in old_data_nodes:
|
|
self.old_data_nodes.discard(n)
|
|
else:
|
|
self.move_data_by_same_tag(xsd_element_node, xml_parent_node, new_nodes)
|
|
|
|
return False
|
|
|
|
def move_data_by_xpath(self, xpath, xsd_element_node, xml_parent_node, new_nodes):
|
|
element_tag = xsd_element_node.get("name")
|
|
|
|
old_data_nodes = self.get_from_old_data(xml_parent_node, xpath)
|
|
if self.complex_type_of_element(xsd_element_node) is None:
|
|
max_occurs_raw = xsd_element_node.get("maxOccurs")
|
|
|
|
# Use `len(old_data_nodes)` to ensure that all old data nodes are moved if an unbound number of
|
|
# occurrences is allowed.
|
|
max_occurs = \
|
|
len(old_data_nodes) if max_occurs_raw == "unbounded" else \
|
|
1 if max_occurs_raw is None else \
|
|
int(max_occurs_raw)
|
|
|
|
if len(old_data_nodes) <= max_occurs:
|
|
for n in old_data_nodes:
|
|
new_node = etree.Element(element_tag)
|
|
new_node.text = n.text
|
|
for k, v in n.items():
|
|
new_node.set(k, v)
|
|
new_nodes.append(new_node)
|
|
self.old_data_nodes.discard(n)
|
|
|
|
return False
|
|
else:
|
|
# For each complex type containing multiple configuration items, this method can only create at most one
|
|
# single node, as there is no way for the default data movers to migrate multiple pieces of data of the same
|
|
# type to the new XML.
|
|
if old_data_nodes:
|
|
new_node = etree.Element(element_tag)
|
|
for k, v in old_data_nodes[0].items():
|
|
new_node.set(k, v)
|
|
new_nodes.append(new_node)
|
|
return True
|
|
|
|
def move_data_by_same_tag(self, xsd_element_node, xml_parent_node, new_nodes):
|
|
element_tag = xsd_element_node.get("name")
|
|
return self.move_data_by_xpath(f".//{element_tag}", xsd_element_node, xml_parent_node, new_nodes)
|
|
|
|
def move_data_null(self, xsd_element_node, xml_parent_node, new_nodes):
|
|
return False
|
|
|
|
data_movers = {
|
|
# Configuration items with the same name but under different parents
|
|
"vm/name": partialmethod(move_data_by_xpath, "./name"),
|
|
"os_config/name": partialmethod(move_data_by_xpath, ".//os_config/name"),
|
|
"epc_section/base": partialmethod(move_data_by_xpath, ".//epc_section/base"),
|
|
"console_vuart/base": partialmethod(move_data_by_xpath, ".//console_vuart/base"),
|
|
"epc_section/size": partialmethod(move_data_by_xpath, ".//epc_section/size"),
|
|
"memory/size": partialmethod(move_data_by_xpath, ".//memory/size"),
|
|
|
|
# Guest flags
|
|
"lapic_passthrough": partialmethod(move_guest_flag, "GUEST_FLAG_LAPIC_PASSTHROUGH"),
|
|
"io_completion_polling": partialmethod(move_guest_flag, "GUEST_FLAG_IO_COMPLETION_POLLING"),
|
|
"nested_virtualization_support": partialmethod(move_guest_flag, "GUEST_FLAG_NVMX_ENABLED"),
|
|
"virtual_cat_support": partialmethod(move_guest_flag, "GUEST_FLAG_VCAT_ENABLED"),
|
|
"secure_world_support": partialmethod(move_guest_flag, "GUEST_FLAG_SECURITY_VM"),
|
|
"hide_mtrr_support": partialmethod(move_guest_flag, "GUEST_FLAG_HIDE_MTRR"),
|
|
"security_vm": partialmethod(move_guest_flag, "GUEST_FLAG_SECURITY_VM"),
|
|
|
|
"legacy_vuart": move_legacy_vuart,
|
|
"vuart_connections": move_vuart_connections,
|
|
"IVSHMEM": move_ivshmem,
|
|
"vm_type": move_vm_type,
|
|
|
|
"default": move_data_by_same_tag,
|
|
}
|
|
|
|
def add_missing_nodes(self, xsd_element_node, xml_parent_node, xml_anchor_node):
|
|
new_nodes = []
|
|
def call_mover(mover):
|
|
if isinstance(mover, partialmethod):
|
|
return mover.__get__(self, type(self))(xsd_element_node, xml_parent_node, new_nodes)
|
|
else:
|
|
return mover(self, xsd_element_node, xml_parent_node, new_nodes)
|
|
|
|
# Common names (such as 'name' or 'base') may be used as tags in multiple places each of which has different
|
|
# meanings. In such cases it is ambiguious to query old data by that common tag alone.
|
|
element_tag = xsd_element_node.get("name")
|
|
element_tag_with_parent = f"{xml_parent_node.tag}/{element_tag}"
|
|
|
|
mover_key = \
|
|
element_tag_with_parent if element_tag_with_parent in self.data_movers.keys() else \
|
|
element_tag if element_tag in self.data_movers.keys() else \
|
|
"default"
|
|
visit_children = call_mover(self.data_movers[mover_key])
|
|
|
|
if xml_anchor_node is not None:
|
|
for n in new_nodes:
|
|
xml_anchor_node.addprevious(n)
|
|
else:
|
|
xml_parent_node.extend(new_nodes)
|
|
|
|
if visit_children:
|
|
return new_nodes
|
|
else:
|
|
return []
|
|
|
|
@property
|
|
@lru_cache
|
|
def upgraded_etree(self):
|
|
new_xml_etree = etree.ElementTree(etree.Element(self.old_xml_etree.getroot().tag))
|
|
root_node = new_xml_etree.getroot()
|
|
for k, v in self.old_xml_etree.getroot().items():
|
|
new_xml_etree.getroot().set(k, v)
|
|
|
|
# Migrate the HV and VM nodes, which are needed to kick off a thorough traversal of the existing scenario.
|
|
for old_node in self.old_xml_etree.getroot():
|
|
new_node = etree.Element(old_node.tag)
|
|
|
|
if old_node.tag == "vm":
|
|
# FIXME: Here we still hard code how the load order of a VM is specified in different versions of
|
|
# schemas. While it is not subject to frequent changes, it would be better if we use a more generic
|
|
# approach instead.
|
|
load_order_node = etree.SubElement(new_node, "load_order")
|
|
|
|
# In the history we have two ways of specifying the load order of a VM: either by vm_type or by
|
|
# loader_order.
|
|
vm_type = old_node.xpath(".//vm_type/text()")
|
|
old_load_order_node = old_node.xpath(".//load_order")
|
|
if old_load_order_node:
|
|
load_order_node.text = old_load_order_node[0].text
|
|
self.old_data_nodes.discard(old_load_order_node[0])
|
|
elif vm_type:
|
|
if vm_type[0].startswith("PRE_") or vm_type[0] in ["SAFETY_VM"]:
|
|
load_order_node.text = "PRE_LAUNCHED_VM"
|
|
elif vm_type[0].startswith("POST_"):
|
|
load_order_node.text = "POST_LAUNCHED_VM"
|
|
else:
|
|
load_order_node.text = "SERVICE_VM"
|
|
else:
|
|
logging.error(f"Cannot infer the loader order of VM {self.old_xml_etree.getelementpath(old_node)}")
|
|
continue
|
|
|
|
root_node.append(new_node)
|
|
for k, v in old_node.items():
|
|
new_node.set(k, v)
|
|
self.hv_vm_node_map[new_node] = old_node
|
|
|
|
# Now fill the rest of configuration items using the old data
|
|
self.transform(new_xml_etree)
|
|
|
|
return new_xml_etree
|
|
|
|
class UpgradingScenarioStage(PipelineStage):
|
|
uses = {"schema_etree", "scenario_etree"}
|
|
provides = {"scenario_etree"}
|
|
|
|
class DiscardedDataFilter(namedtuple("DiscardedDataFilter", ["path", "data", "info"])):
|
|
def filter(self, path, data):
|
|
simp_path = re.sub(r"\[[^\]]*\]", "", path)
|
|
if not simp_path.endswith(self.path):
|
|
return False
|
|
if self.data and data != self.data:
|
|
return False
|
|
|
|
if self.info:
|
|
logging.info(f"{path} = '{data}': {self.info}")
|
|
return True
|
|
|
|
filters = [
|
|
DiscardedDataFilter("hv/FEATURES/IVSHMEM", None, "IVSHMEM is now automatically enabled if any IVSHMEM region is specified."),
|
|
DiscardedDataFilter("hv/FEATURES/NVMX_ENABLED", None, "Nest virtualization support is now automatically included if enabled for any VM."),
|
|
DiscardedDataFilter("hv/MISC_CFG/UEFI_OS_LOADER_NAME", None, None),
|
|
DiscardedDataFilter("vm/guest_flags/guest_flag", "0", None),
|
|
DiscardedDataFilter("vm/epc_section/base", "0", "Post-launched VMs cannot have EPC sections."),
|
|
DiscardedDataFilter("vm/epc_section/size", "0", "Post-launched VMs cannot have EPC sections."),
|
|
DiscardedDataFilter("vm/cpu_affinity/pcpu_id", None, "CPU affinity of the service VM is no longer needed."),
|
|
DiscardedDataFilter("vm/os_config/name", None, "Guest OS names are no longer needed in scenario definitions."),
|
|
]
|
|
|
|
def run(self, obj):
|
|
upgrader = ScenarioUpgrader(obj.get("schema_etree"), obj.get("scenario_etree"))
|
|
new_scenario_etree = upgrader.upgraded_etree
|
|
|
|
discarded_data = [(obj.get("scenario_etree").getelementpath(n), n.text) for n in upgrader.old_data_nodes]
|
|
for path, data in sorted(discarded_data):
|
|
if not any(map(lambda x: x.filter(path, data), self.filters)):
|
|
escaped_data = data.replace("\n", "\\n")
|
|
logging.warning(f"{path} = '{escaped_data}' is discarded")
|
|
|
|
obj.set("scenario_etree", new_scenario_etree)
|
|
|
|
def main(args):
|
|
pipeline = PipelineEngine(["schema_path", "scenario_path"])
|
|
pipeline.add_stages([
|
|
LXMLLoadStage("schema"),
|
|
LXMLLoadStage("scenario"),
|
|
SlicingSchemaByVMTypeStage(),
|
|
UpgradingScenarioStage(),
|
|
])
|
|
|
|
obj = PipelineObject(schema_path = args.schema, scenario_path = args.scenario)
|
|
pipeline.run(obj)
|
|
# We know we are using lxml to parse the scenario XML, so it is ok to use lxml specific write options here.
|
|
obj.get("scenario_etree").write(args.out, pretty_print=True)
|
|
|
|
if __name__ == "__main__":
|
|
config_tools_dir = os.path.join(os.path.dirname(__file__), "..")
|
|
schema_dir = os.path.join(config_tools_dir, "schema")
|
|
|
|
parser = argparse.ArgumentParser(description="Try adapting data in a scenario XML to the latest schema.")
|
|
parser.add_argument("scenario", help="Path to the scenario XML file from users")
|
|
parser.add_argument("out", nargs="?", default="out.xml", help="Path where the output is placed")
|
|
parser.add_argument("--schema", default=os.path.join(schema_dir, "config.xsd"), help="the XML schema that defines the syntax of scenario XMLs")
|
|
args = parser.parse_args()
|
|
|
|
logging.basicConfig(level="INFO")
|
|
main(args)
|