# This Source Code Form is subject to the terms of the Mozilla Public # License, v. 2.0. If a copy of the MPL was not distributed with this # file, You can obtain one at http://mozilla.org/MPL/2.0/.
import datetime import errno import json import os import posixpath import re import subprocess from collections import defaultdict from typing import Dict
import mozpack.path as mozpath import requests import six.moves.urllib_parse as urlparse from mozbuild.base import MachCommandConditions as conditions from mozbuild.base import MozbuildObject from mozfile import which from mozinfo.platforminfo import PlatformInfo from moztest.resolve import TestManifestLoader, TestResolver from redo import retriable
self._hg = None if conditions.is_hg(self.build_obj):
self._hg = which("hg") ifnot self._hg: raise OSError(errno.ENOENT, "Could not find 'hg' on PATH.")
self._git = None if conditions.is_git(self.build_obj):
self._git = which("git") ifnot self._git: raise OSError(errno.ENOENT, "Could not find 'git' on PATH.")
def find_in_hg_or_git(self, test_name): if self._hg:
cmd = [self._hg, "files", "-I", test_name] elif self._git:
cmd = [self._git, "ls-files", test_name] else: returnNone try:
out = subprocess.check_output(cmd, universal_newlines=True).splitlines() except subprocess.CalledProcessError:
out = None return out
def set_test_name(self): # Generating a unified report for a specific test is complicated # by differences in the test name used in various data sources. # Consider: # - It is often convenient to request a report based only on # a short file name, rather than the full path; # - Bugs may be filed in bugzilla against a simple, short test # name or the full path to the test; # This function attempts to find appropriate names for different # queries based on the specified test name.
# full_test_name is full path to file in hg (or git)
self.full_test_name = None
out = self.find_in_hg_or_git(self.test_name) if out and len(out) == 1:
self.full_test_name = out[0] elif out and len(out) > 1:
print("Ambiguous test name specified. Found:") for line in out:
print(line) else:
out = self.find_in_hg_or_git("**/%s*" % self.test_name) if out and len(out) == 1:
self.full_test_name = out[0] elif out and len(out) > 1:
print("Ambiguous test name. Found:") for line in out:
print(line) if self.full_test_name:
self.full_test_name.replace(os.sep, posixpath.sep)
print("Found %s in source control." % self.full_test_name) else:
print("Unable to validate test name '%s'!" % self.test_name)
self.full_test_name = self.test_name
# search for full_test_name in test manifests
here = os.path.abspath(os.path.dirname(__file__))
resolver = TestResolver.from_environment(
cwd=here, loader_cls=TestManifestLoader
)
relpath = self.build_obj._wrap_path_argument(self.full_test_name).relpath()
tests = list(resolver.resolve_tests(paths=[relpath])) if len(tests) == 1:
relpath = self.build_obj._wrap_path_argument(tests[0]["manifest"]).relpath()
print("%s found in manifest %s" % (self.full_test_name, relpath)) if tests[0].get("flavor"):
print(" flavor: %s" % tests[0]["flavor"]) if tests[0].get("skip-if"):
print(" skip-if: %s" % tests[0]["skip-if"]) if tests[0].get("fail-if"):
print(" fail-if: %s" % tests[0]["fail-if"]) elif len(tests) == 0:
print("%s not found in any test manifest!" % self.full_test_name) else:
print("%s found in more than one manifest!" % self.full_test_name)
# short_name is full_test_name without path
self.short_name = None
name_idx = self.full_test_name.rfind("/") if name_idx > 0:
self.short_name = self.full_test_name[name_idx + 1 :] if self.short_name and self.short_name == self.test_name:
self.short_name = None
def get_platform(self, record): if"platform"in record["build"]:
platform = record["build"]["platform"] else:
platform = "-"
platform_words = platform.split("-")
types_label = "" # combine run and build types and eliminate duplicates
run_types = [] if"run"in record and"type"in record["run"]:
run_types = record["run"]["type"]
run_types = run_types if isinstance(run_types, list) else [run_types]
build_types = [] if"build"in record and"type"in record["build"]:
build_types = record["build"]["type"]
build_types = (
build_types if isinstance(build_types, list) else [build_types]
)
run_types = list(set(run_types + build_types)) # '1proc' is used as a treeherder label but does not appear in run types if"e10s"notin run_types:
run_types = run_types + ["1proc"] for run_type in run_types: # chunked is not interesting if run_type == "chunked": continue # e10s is the default: implied if run_type == "e10s": continue # sometimes a build/run type is already present in the build platform if run_type in platform_words: continue if types_label:
types_label += "-"
types_label += run_type return"%s/%s:" % (platform, types_label)
def report_bugs(self): # Report open bugs matching test name
search = self.full_test_name if self.test_name:
search = "%s,%s" % (search, self.test_name) if self.short_name:
search = "%s,%s" % (search, self.short_name)
payload = {"quicksearch": search, "include_fields": "id,summary"}
response = requests.get("https://bugzilla.mozilla.org/rest/bug", payload)
response.raise_for_status()
json_response = response.json()
print("\nBugzilla quick search for '%s':" % search) if"bugs"in json_response: for bug in json_response["bugs"]:
print("Bug %s: %s" % (bug["id"], bug["summary"])) else:
print("No bugs found.")
ifnot self.show_info andnot show_bugs: # by default, show everything
self.show_info = True
show_bugs = True
for test_name in test_names:
print("===== %s =====" % test_name)
self.test_name = test_name if len(self.test_name) < 6:
print("'%s' is too short for a test name!" % self.test_name) continue
self.set_test_name() if show_bugs:
self.report_bugs()
class TestInfoReport(TestInfo): """
Support 'mach test-info report': Report of test runs summarized by
manifest and component. """
@retriable(attempts=3, sleeptime=5, sleepscale=2) def get_url(self, target_url): # if we fail to get valid json (i.e. end point has malformed data), return {}
retVal = {} try:
self.log_verbose("getting url: %s" % target_url)
r = requests.get(target_url, headers={"User-agent": "mach-test-info/1.0"})
self.log_verbose("got status: %s" % r.status_code)
r.raise_for_status()
retVal = r.json() except json.decoder.JSONDecodeError:
self.log_verbose("Error retrieving data from %s" % target_url)
return retVal
def update_report(self, by_component, result, path_mod): def update_item(item, label, value): # It is important to include any existing item value in case ActiveData # returns multiple records for the same test; that can happen if the report # sometimes maps more than one ActiveData record to the same path.
new_value = item.get(label, 0) + value if type(new_value) is int:
item[label] = new_value else:
item[label] = float(round(new_value, 2)) # pylint: disable=W1633
if"test"in result and"tests"in by_component:
test = result["test"] if path_mod:
test = path_mod(test) for bc in by_component["tests"]: for item in by_component["tests"][bc]: if test == item["test"]: # pylint: disable=W1633
seconds = float(round(result.get("duration", 0), 2))
update_item(item, "total run time, seconds", seconds)
update_item(item, "total runs", result.get("count", 0))
update_item(item, "skipped runs", result.get("skips", 0))
update_item(item, "failed runs", result.get("failures", 0)) returnTrue returnFalse
def description(
self,
components,
flavor,
subsuite,
paths,
show_manifests,
show_tests,
show_summary,
show_annotations,
filter_values,
filter_keys,
start_date,
end_date,
): # provide a natural language description of the report options
what = [] if show_manifests:
what.append("test manifests") if show_tests:
what.append("tests") if show_annotations:
what.append("test manifest annotations") if show_summary and len(what) == 0:
what.append("summary of tests only") if len(what) > 1:
what[-1] = "and " + what[-1]
what = ", ".join(what)
d = "Test summary report for " + what if components:
d += ", in specified components (%s)" % components else:
d += ", in all components" if flavor:
d += ", in specified flavor (%s)" % flavor if subsuite:
d += ", in specified subsuite (%s)" % subsuite if paths:
d += ", on specified paths (%s)" % paths if filter_values:
d += ", containing '%s'" % filter_values if filter_keys:
d += " in manifest keys '%s'" % filter_keys else:
d += " in any part of manifest entry"
d += ", including historical run-time data for the last "
start = datetime.datetime.strptime(start_date, "%Y-%m-%d")
end = datetime.datetime.strptime(end_date, "%Y-%m-%d")
d += "%s days on trunk (autoland/m-c)" % ((end - start).days)
d += " as of %s." % end_date return d
# TODO: this is hacked for now and very limited def parse_test(self, summary): if summary.endswith("single tracking bug"):
name_part = summary.split("|")[0] # remove 'single tracking bug'
name_part.strip() return name_part.split()[-1] # get just the test name, not extra words returnNone
def get_runcount_data(self, runcounts_input_file, start, end): # TODO: use start/end properly if runcounts_input_file: try: with open(runcounts_input_file, "r") as f:
runcounts = json.load(f) except:
print("Unable to load runcounts from path: %s" % runcounts_input_file) raise else:
runcounts = self.get_runcounts(days=MAX_DAYS)
runcounts = self.squash_runcounts(runcounts, days=MAX_DAYS) return runcounts
url = "" for artifact in artifacts: if artifact["name"].endswith("test-run-info.json"):
url = queue.buildUrl("getLatestArtifact", task_id, artifact["name"]) break return url
def get_runcounts(self, days=MAX_DAYS):
testrundata = {} # get historical data from test-info job artifact; if missing get fresh
url = self.get_testinfoall_index_url()
print("INFO: requesting runcounts url: %s" % url)
olddata = self.get_url(url)
# fill in any holes we have
endday = datetime.datetime.now(datetime.timezone.utc) - datetime.timedelta(
days=1
)
startday = endday - datetime.timedelta(days=days)
urls_to_fetch = [] # build list of dates with missing data while startday < endday:
nextday = startday + datetime.timedelta(days=1) ifnot olddata.get(str(nextday.date()), {}):
url = "https://treeherder.mozilla.org/api/groupsummary/"
url += "?startdate=%s&enddate=%s" % (
startday.date(),
nextday.date(),
)
urls_to_fetch.append([str(nextday.date()), url])
testrundata[str(nextday.date())] = olddata.get(str(nextday.date()), {})
startday = nextday
# limit missing data collection to 5 most recent days days to reduce overall runtime for date, url in urls_to_fetch[-5:]: try:
testrundata[date] = self.get_url(url) except requests.exceptions.HTTPError: # We want to see other errors, but can accept HTTPError failures
print(f"Unable to retrieve results for url: {url}") pass
return testrundata
def squash_runcounts(self, runcounts, days=MAX_DAYS): # squash all testrundata together into 1 big happy family for the last X days
endday = datetime.datetime.now(datetime.timezone.utc) - datetime.timedelta(
days=1
)
oldest = endday - datetime.timedelta(days=days)
retVal = {} for datekey in runcounts.keys(): # strip out older days if datetime.date.fromisoformat(datekey) < oldest.date(): continue
jtn = runcounts[datekey].get("job_type_names", {}) ifnot jtn:
print("Warning: Missing job type names from date: %s" % datekey) continue
# TODO: changed this to include all manifests, not just first for m in runcounts[datekey]["manifests"]: for man_name in m.keys(): for job_type_id, result, classification, count in m[man_name]: # HACK: we treat shippable and opt the same for mozinfo and runcounts
job_name = jtn[job_type_id].replace("-shippable", "")
# get bug data for summary, 800 bugs at a time # i.e. https://b.m.o/rest/bug?include_fields=id,product,component,summary&id=1,2,3...
max_bugs = 800
bug_data = []
fields = ["id", "product", "component", "summary"] for bug_index in range(0, len(buglist), max_bugs):
bugs = [str(x) for x in buglist[bug_index : bug_index + max_bugs]] ifnot bugs:
print(f"warning: found no bugs in range {bug_index}, +{max_bugs}") continue
# for each summary, parse filename, store component # IF we find >1 bug with same testname, for now summarize as one for bug in bug_data:
test_name = self.parse_test(bug["summary"]) ifnot test_name: continue
c = int([x["bug_count"] for x in if_data if x["bug_id"] == bug["id"]][0]) if test_name notin retVal.keys():
retVal[test_name] = { "id": bug["id"], "count": 0, "product": bug["product"], "component": bug["component"],
}
retVal[test_name]["count"] += c
def report(
self,
components,
flavor,
subsuite,
paths,
show_manifests,
show_tests,
show_summary,
show_annotations,
filter_values,
filter_keys,
show_components,
output_file,
start,
end,
show_testruns,
runcounts_input_file,
config_matrix_output_file,
): def matches_filters(test): """ ReturnTrueif all of the requested filter_values are found in this test; if filter_keys are specified, restrict search to those test keys. """ for value in filter_values:
value_found = False for key in test: ifnot filter_keys or key in filter_keys: if re.search(value, test[key]):
value_found = True break ifnot value_found: returnFalse returnTrue
start_time = datetime.datetime.now()
# Ensure useful report by default if ( not show_manifests andnot show_tests andnot show_summary andnot show_annotations
):
show_manifests = True
show_summary = True
by_component = {} if components:
components = components.split(",") if filter_keys:
filter_keys = filter_keys.split(",") if filter_values:
filter_values = filter_values.split(",") else:
filter_values = []
display_keys = (filter_keys or []) + ["skip-if", "fail-if", "fails-if"]
display_keys = set(display_keys)
ifd = self.get_intermittent_failure_data(start, end)
manifest_paths = set() for t in tests: if t.get("manifest", None):
manifest_path = t["manifest"] if t.get("ancestor_manifest", None):
manifest_path = "%s:%s" % (t["ancestor_manifest"], t["manifest"])
manifest_paths.add(manifest_path)
manifest_count = len(manifest_paths)
print( "Resolver found {} tests, {} manifests".format(len(tests), manifest_count)
)
if config_matrix_output_file:
topsrcdir = self.build_obj.topsrcdir
config_matrix = {} for manifest in manifest_paths: # we want the first part of the parent:child, as parent shows up in MHTP # TODO: figure out a better solution for child manifests if".toml"in manifest:
relpath = mozpath.relpath(
f"{manifest.split('.toml')[0]}.toml", topsrcdir
) else:
relpath = mozpath.relpath(manifest, topsrcdir) # hack for wpt manifests if relpath.startswith(".."):
relpath = "/" + relpath.replace("../", "")
config_matrix[relpath] = self.create_matrix_from_task_graph(
relpath, runcount
)
self.write_report(config_matrix, config_matrix_output_file)
if show_manifests:
topsrcdir = self.build_obj.topsrcdir
by_component["manifests"] = {}
manifest_paths = list(manifest_paths)
manifest_paths.sort()
relpaths = [] for manifest_path in manifest_paths:
relpath = mozpath.relpath(manifest_path, topsrcdir) if mozpath.commonprefix((manifest_path, topsrcdir)) != topsrcdir: continue
relpaths.append(relpath)
reader = self.build_obj.mozbuild_reader(config_mode="empty")
files_info = reader.files_info(relpaths) for manifest_path in manifest_paths:
relpath = mozpath.relpath(manifest_path, topsrcdir) if mozpath.commonprefix((manifest_path, topsrcdir)) != topsrcdir: continue
manifest_info = None if relpath in files_info:
bug_component = files_info[relpath].get("BUG_COMPONENT") if bug_component:
key = "{}::{}".format(
bug_component.product, bug_component.component
) else:
key = "" if (not components) or (key in components):
manifest_info = {"manifest": relpath, "tests": 0, "skipped": 0}
rkey = key if show_components else"all" if rkey in by_component["manifests"]:
by_component["manifests"][rkey].append(manifest_info) else:
by_component["manifests"][rkey] = [manifest_info] if manifest_info: for t in tests: if t["manifest"] == manifest_path:
manifest_info["tests"] += 1 if t.get("skip-if"):
manifest_info["skipped"] += 1 for key in by_component["manifests"]:
by_component["manifests"][key].sort(key=lambda k: k["manifest"])
if show_tests:
by_component["tests"] = {}
if show_tests or show_summary or show_annotations:
test_count = 0
failed_count = 0
skipped_count = 0
annotation_count = 0
condition_count = 0
component_set = set()
relpaths = []
conditions = {}
known_unconditional_annotations = ["skip", "fail", "asserts", "random"]
known_conditional_annotations = [ "skip-if", "fail-if", "run-if", "fails-if", "fuzzy-if", "random-if", "asserts-if",
] for t in tests:
relpath = t.get("srcdir_relpath")
relpaths.append(relpath)
reader = self.build_obj.mozbuild_reader(config_mode="empty")
files_info = reader.files_info(relpaths) for t in tests: ifnot matches_filters(t): continue if"referenced-test"in t: # Avoid double-counting reftests: disregard reference file entries continue if show_annotations: for key in t: if key in known_unconditional_annotations:
annotation_count += 1 if key in known_conditional_annotations:
annotation_count += 1 # Here 'key' is a manifest annotation type like 'skip-if' and t[key] # is the associated condition. For example, the manifestparser # manifest annotation, "skip-if = os == 'win'", is expected to be # encoded as t['skip-if'] = "os == 'win'". # To allow for reftest manifests, t[key] may have multiple entries # separated by ';', each corresponding to a condition for that test # and annotation type. For example, # "skip-if(Android&&webrender) skip-if(OSX)", would be # encoded as t['skip-if'] = "Android&&webrender;OSX".
annotation_conditions = t[key].split(";")
# if key has \n in it, we need to strip it. for manifestparser format # 1) from the beginning of the line # 2) different conditions if in the middle of the line
annotation_conditions = [
x.strip("\n") for x in annotation_conditions
]
temp = [] for condition in annotation_conditions:
temp.extend(condition.split("\n"))
annotation_conditions = temp
for c in annotation_conditions:
condition_count += 1 # Trim reftest fuzzy-if ranges: everything after the first comma # eg. "Android,0-2,1-3" -> "Android"
condition = c.split(",")[0] if condition notin conditions:
conditions[condition] = 0
conditions[condition] += 1
test_count += 1
relpath = t.get("srcdir_relpath") if relpath in files_info:
bug_component = files_info[relpath].get("BUG_COMPONENT") if bug_component:
key = "{}::{}".format(
bug_component.product, bug_component.component
) else:
key = "" if (not components) or (key in components):
component_set.add(key)
test_info = {"test": relpath} for test_key in display_keys:
value = t.get(test_key) if value:
test_info[test_key] = value if t.get("fail-if"):
failed_count += 1 if t.get("fails-if"):
failed_count += 1 if t.get("skip-if"):
skipped_count += 1
# handle included manifests as ancestor:child if t.get("ancestor_manifest", None):
test_info["manifest"] = [ "%s:%s"
% (t["ancestor_manifest"], test_info["manifest"][0])
]
# add in intermittent failure data if ifd.get(relpath):
if_data = ifd.get(relpath)
test_info["failure_count"] = if_data["count"] if show_testruns:
total_runs = 0 for m in test_info["manifest"]: if m in runcount.keys(): for x in runcount.get(m, []): ifnot x: break
total_runs += x[3] if total_runs > 0:
test_info["total_runs"] = total_runs
if show_tests:
rkey = key if show_components else"all" if rkey in by_component["tests"]: # Avoid duplicates: Some test paths have multiple TestResolver # entries, as when a test is included by multiple manifests.
found = False for ctest in by_component["tests"][rkey]: if ctest["test"] == test_info["test"]:
found = True break ifnot found:
by_component["tests"][rkey].append(test_info) else: for ti in by_component["tests"][rkey]: if ti["test"] == test_info["test"]: if (
test_info["manifest"][0] notin ti["manifest"]
):
ti_manifest = test_info["manifest"] if test_info.get( "ancestor_manifest", None
):
ti_manifest = "%s:%s" % (
test_info["ancestor_manifest"],
ti_manifest,
)
ti["manifest"].extend(ti_manifest) else:
by_component["tests"][rkey] = [test_info] if show_tests: for key in by_component["tests"]:
by_component["tests"][key].sort(key=lambda k: k["test"])
def diff_summaries(self, by_component, summary1, summary2): """
Update by_component with comparison of summaries. """
all_keys = set(summary1.keys()) | set(summary2.keys()) for key in all_keys:
delta = summary2.get(key, 0) - summary1.get(key, 0)
by_component["summary"]["%s delta" % key] = delta
def diff_all_components(self, by_component, tests1, tests2): """
Update by_component with any added/deleted tests, for all components. """
self.added_count = 0
self.deleted_count = 0 for component in tests1:
component1 = tests1[component]
component2 = [] if component notin tests2 else tests2[component]
self.diff_component(by_component, component, component1, component2) for component in tests2: if component notin tests1:
component2 = tests2[component]
self.diff_component(by_component, component, [], component2)
by_component["summary"]["added tests"] = self.added_count
by_component["summary"]["deleted tests"] = self.deleted_count
def diff_component(self, by_component, component, component1, component2): """
Update by_component[component] with any added/deleted tests for the
named component. "added": tests found in component2 but missing from component1. "deleted": tests found in component1 but missing from component2. """
tests1 = set([t["test"] for t in component1])
tests2 = set([t["test"] for t in component2])
deleted = tests1 - tests2
added = tests2 - tests1 if deleted or added:
by_component["tests"][component] = {} if deleted:
by_component["tests"][component]["deleted"] = sorted(list(deleted)) if added:
by_component["tests"][component]["added"] = sorted(list(added))
self.added_count += len(added)
self.deleted_count += len(deleted)
common = len(tests1.intersection(tests2))
self.log_verbose( "%s: %d deleted, %d added, %d common"
% (component, len(deleted), len(added), common)
)
################################################################################ ### ### Below is code for creating a os/version/processor/config/variant matrix ###
response = requests.get(url, headers={"User-agent": "mach-test-info/1.0"})
data = response.json() with open(filename, "w") as f:
json.dump(data, f)
for task in data.values():
task_label: str = task["label"] # HACK: we treat shippable and opt the same from mozinfo, runtime counts
task_label = task_label.replace("-shippable", "")
# we only want test tasks ifnot task_label.startswith("test-"): continue if task_label.endswith("-cf"): continue
try:
parts = task_label.split("-") if int(parts[-1]):
task_label = "-".join(parts[:-1]) except ValueError: pass
# TODO: this only works for tasks where we schedule by manifest
env = task.get("task", {}).get("payload", {}).get("env", {})
mhtp = json.loads(env.get("MOZHARNESS_TEST_PATHS", "{}")) ifnot mhtp: # mock up logic here if matching task
suite = self.find_non_test_path_loader(task_label) ifnot suite: continue
mhtp[suite] = [suite]
# TODO: figure out a better method for dealing with TEST_TAG # when we have a test_tag, all skipped manifests are added to chunk 1. # we are skipping real manifests, but avoiding many overreported manifests. # # NOTE: some variants only have a single chunk, so no numbers if json.loads(env.get("MOZHARNESS_TEST_TAG", "{}")): ifnot json.loads(env.get("MOZHARNESS_TEST_PATHS", "{}")): # mock up logic here if matching task
suite = self.find_non_test_path_loader(task_label) ifnot suite: continue
mhtp[suite] = [suite]
for suite in mhtp: for manifest in mhtp[suite]:
self.matrix_map[manifest].append(task_label)
extra = task.get("task", {}).get("extra", {}).get("test-setting", {})
platform_info = PlatformInfo(extra)
def find_non_test_path_loader(self, label): # TODO: how to keep this list synchronized?
known_suites = [ "mochitest-browser-media", "telemetry-tests-client", "mochitest-webgl2-ext", "mochitest-webgl1-ext", "jittest-1proc", "mochitest-browser-translations", "jsreftest", "mochitest-browser-screenshots", "marionette-unittest",
]
match = [x for x in known_suites if x in label] if match: return match[0] return""
# find manifest in matrix_map and for all tasks that run this # pull the tuples out and create a definitive list def create_matrix_from_task_graph(self, target_manifest, runcount):
results = {}
ifnot self.matrix_map:
self.build_matrix_cache()
# for tasks with no MOZHARNESS_TEST_PATHS, provide basic data if target_manifest in runcount and self.find_non_test_path_loader(
runcount[target_manifest][0][0]
):
suite = self.find_non_test_path_loader(runcount[target_manifest][0][0])
self.matrix_map[target_manifest] = self.matrix_map[suite]
for tl in self.matrix_map.get(target_manifest, []):
task_label = tl.replace("-shippable", "")
platform_info = self.task_tuples[task_label]
# add in runcounts, we can find find the index of the given task_label in 'job_type_names', # use that to get specific runs
passed = 0
failed = 0 if target_manifest in runcount: # data = [[job_name, result, classification, count], ...] for data in [
x for x in runcount[target_manifest] if task_label == x[0]
]: if data[1] == "passed":
passed += data[-1] else:
failed += data[-1]
# this helps avoid 'skipped' manifests if passed == 0 and failed == 0: continue
if platform_info.os notin results:
results[platform_info.os] = {}
os = results[platform_info.os] if platform_info.os_version notin os:
os[platform_info.os_version] = {}
os_version = os[platform_info.os_version] if platform_info.arch notin os_version:
os_version[platform_info.arch] = {}
arch = os_version[platform_info.arch] if platform_info.build_type notin arch:
arch[platform_info.build_type] = {}
Die Informationen auf dieser Webseite wurden
nach bestem Wissen sorgfältig zusammengestellt. Es wird jedoch weder Vollständigkeit, noch Richtigkeit,
noch Qualität der bereit gestellten Informationen zugesichert.
Bemerkung:
Die farbliche Syntaxdarstellung ist noch experimentell.