#!/usr/bin/python3
# Generate testcase files and Makefile fragments for DSO sorting test
# Copyright (C) 2021-2022 Free Software Foundation, Inc.
# This file is part of the GNU C Library.
#
# The GNU C Library is free software; you can redistribute it and/or
# modify it under the terms of the GNU Lesser General Public
# License as published by the Free Software Foundation; either
# version 2.1 of the License, or (at your option) any later version.
#
# The GNU C Library is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
# Lesser General Public License for more details.
#
# You should have received a copy of the GNU Lesser General Public
# License along with the GNU C Library; if not, see
# .
"""Generate testcase files and Makefile fragments for DSO sorting test
This script takes a small description string language, and generates
testcases for displaying the ELF dynamic linker's dependency sorting
behavior, allowing verification.
Testcase descriptions are semicolon-separated description strings, and
this tool generates a testcase from the description, including main program,
associated modules, and Makefile fragments for including into elf/Makefile.
This allows automation of what otherwise would be very laborous manual
construction of complex dependency cases, however it must be noted that this
is only a tool to speed up testcase construction, and thus the generation
features are largely mechanical in nature; inconsistencies or errors may occur
if the input description was itself erroneous or have unforeseen interactions.
The format of the input test description files are:
# Each test description has a name, lines of description,
# and an expected output specification. Comments use '#'.
testname1:
output:
# Tests can be marked to be XFAIL by using 'xfail_output' instead
testname2:
xfail_output:
# A default set of GLIBC_TUNABLES tunables can be specified, for which
# all following tests will run multiple times, once for each of the
# GLIBC_TUNABLES=... strings set by the 'tunable_option' command.
tunable_option:
tunable_option:
# Test descriptions can use multiple lines, which will all be merged
# together, so order is not important.
testname3:
...
output:
# 'testname3' will be run and compared two times, for both
# GLIBC_TUNABLES= and
# GLIBC_TUNABLES=. This can be cleared and reset by the
# 'clear_tunables' command:
clear_tunables
# Multiple expected outputs can also be specified, with an associated
# tunable option in (), which multiple tests will be run with each
# GLIBC_TUNABLES=... option tried.
testname4:
...
output():
output():
# Individual tunable output cases can be XFAILed, though note that
# this will have the effect of XFAILing the entire 'testname4' test
# in the final top-level tests.sum summary.
xfail_output():
# When multiple outputs (with specific tunable strings) are specified,
# these take priority over any active 'tunable_option' settings.
# When a test is meant to be placed under 'xtests' (not run under
# "make check", but only when "make xtests" is used), the testcase name can be
# declared using 'xtest()':
...
xtest(test-too-big1):
output:
...
# Do note that under current elf/Makefile organization, for such a xtest case,
# while the test execution is only run under 'make xtests', the associated
# DSOs are always built even under 'make check'.
On the description language used, an example description line string:
a->b!->[cdef];c=>g=>h;{+c;%c;-c}->a
Each identifier represents a shared object module, currently sequences of
letters/digits are allowed, case-sensitive.
All such shared objects have a constructor/destructor generated for them
that emits its name followed by a '>' for constructors, and '<' followed by
its name for destructors, e.g. if the name is 'obj1', then "obj1>" and " operator specifies a link time dependency, these can be chained for
convenience (e.g. a->b->c->d).
The => operator creates a call-reference, e.g. for a=>b, an fn_a() function
is created inside module 'a', which calls fn_b() in module 'b'.
These module functions emit 'name()' output in nested form,
e.g. a=>b emits 'a(b())'
For single character object names, square brackets [] in the description
allows specifying multiple objects; e.g. a->[bcd]->e is equivalent to
a->b->e;a->c->e;a->d->e
The () parenthesis construct with space separated names is also allowed for
specifying objects. For names with integer suffixes a range can also be used,
e.g. (foo1 bar2-5), specifies DSOs foo1, bar2, bar2, bar3, bar4, bar5.
A {} construct specifies the main test program, and its link dependencies
are also specified using ->. Inside {}, a few ;-separated constructs are
allowed:
+a Loads module a using dlopen(RTLD_LAZY|RTLD_GLOBAL)
^a Loads module a using dlopen(RTLD_LAZY)
%a Use dlsym() to load and call fn_a()
@a Calls fn_a() directly.
-a Unloads module a using dlclose()
The generated main program outputs '{' '}' with all output from above
constructs in between. The other output before/after {} are the ordered
constructor/destructor output.
If no {} construct is present, a default empty main program is linked
against all objects which have no dependency linked to it. e.g. for
'[ab]->c;d->e', the default main program is equivalent to '{}->[abd]'
Sometimes for very complex or large testcases, besides specifying a
few explicit dependencies from main{}, the above default dependency
behavior is still useful to automatically have, but is turned off
upon specifying a single explicit {}->dso_name.
In this case, add {}->* to explicitly add this generation behavior:
# Main program links to 'foo', and all other objects which have no
# dependency linked to it.
{}->foo,{}->*
Note that '*' works not only on main{}, but can be used as the
dependency target of any object. Note that it only works as a target,
not a dependency source.
The '!' operator after object names turns on permutation of its
dependencies, e.g. while a->[bcd] only generates one set of objects,
with 'a.so' built with a link line of "b.so c.so d.so", for a!->[bcd]
permutations of a's dependencies creates multiple testcases with
different link line orders: "b.so c.so d.so", "c.so b.so d.so",
"b.so d.so c.so", etc. Note that for a specified on
the script command-line, multiple , , etc.
tests will be generated (e.g. for a!->[bc]!->[de], eight tests with
different link orders for a, b, and c will be generated)
It is possible to specify the ELF soname field for an object or the
main program:
# DSO 'a' will be linked with the appropriate -Wl,-soname=x setting
a->b->c;soname(a)=x
# The the main program can also have a soname specified
soname({})=y
This can be used to test how ld.so behaves when objects and/or the
main program have such a field set.
Strings Output by Generated Testcase Programs
The text output produced by a generated testcase consists of three main
parts:
1. The constructors' output
2. Output from the main program
3. Destructors' output
To see by example, a simple test description "a->b->c" generates a testcase
that when run, outputs: "c>b>a>{}' character,
and the "c>b>a" part above is the full constructor output by all DSOs, the
order indicating that DSO 'c', which does not depend on any other DSO, has
its constructor run first, followed by 'b' and then 'a'.
Destructor output for each DSO is a '<' character followed by its name,
reflecting its reverse nature of constructors. In the above example, the
destructor output part is "g=>h;{+c;%c;-c}->a->h
This produces a testcase, that when executed outputs:
h>a>{+c[g>c>];%c();-c[h dependency as expected.
Inside the main program, the "+c" action triggers a dlopen() of DSO 'c',
causing another chain of constructors "g>c>" to be triggered. Here it is
displayed inside [] brackets for each dlopen call. The same is done for "-c",
a dlclose() of 'c'.
The "%c" output is due to calling to fn_c() inside DSO 'c', this comprises
of two parts: the '%' character is printed by the caller, here it is the main
program. The 'c' character is printed from inside fn_c(). The '%' character
indicates that this is called by a dlsym() of "fn_c". A '@' character would
mean a direct call (with a symbol reference). These can all be controlled
by the main test program constructs documented earlier.
The output strings described here is the exact same form placed in
test description files' "output: " line.
"""
import sys
import re
import os
import subprocess
import argparse
from collections import OrderedDict
import itertools
# BUILD_GCC is only used under the --build option,
# which builds the generated testcase, including DSOs using BUILD_GCC.
# Mainly for testing purposes, especially debugging of this script,
# and can be changed here to another toolchain path if needed.
build_gcc = "gcc"
def get_parser():
parser = argparse.ArgumentParser("")
parser.add_argument("description",
help="Description string of DSO dependency test to be "
"generated (see script source for documentation of "
"description language), either specified here as "
"command line argument, or by input file using "
"-f/--description-file option",
nargs="?", default="")
parser.add_argument("test_name",
help="Identifier for testcase being generated",
nargs="?", default="")
parser.add_argument("--objpfx",
help="Path to place generated files, defaults to "
"current directory if none specified",
nargs="?", default="./")
parser.add_argument("-m", "--output-makefile",
help="File to write Makefile fragment to, defaults to "
"stdout when option not present",
nargs="?", default="")
parser.add_argument("-f", "--description-file",
help="Input file containing testcase descriptions",
nargs="?", default="")
parser.add_argument("--build", help="After C testcase generated, build it "
"using gcc (for manual testing purposes)",
action="store_true")
parser.add_argument("--debug-output",
help="Prints some internal data "
"structures; used for debugging of this script",
action="store_true")
return parser
# Main script starts here.
cmdlineargs = get_parser().parse_args()
test_name = cmdlineargs.test_name
description = cmdlineargs.description
objpfx = cmdlineargs.objpfx
description_file = cmdlineargs.description_file
output_makefile = cmdlineargs.output_makefile
makefile = ""
default_tunable_options = []
current_input_lineno = 0
def error(msg):
global current_input_lineno
print("Error: %s%s" % ((("Line %d, " % current_input_lineno)
if current_input_lineno != 0 else ""),
msg))
exit(1)
if(test_name or description) and description_file:
error("both command-line testcase and input file specified")
if test_name and not description:
error("command-line testcase name without description string")
# Main class type describing a testcase.
class TestDescr:
def __init__(self):
self.objs = [] # list of all DSO objects
self.deps = OrderedDict() # map of DSO object -> list of dependencies
# map of DSO object -> list of call refs
self.callrefs = OrderedDict()
# map of DSO object -> list of permutations of dependencies
self.dep_permutations = OrderedDict()
# map of DSO object -> SONAME of object (if one is specified)
self.soname_map = OrderedDict()
# list of main program operations
self.main_program = []
# set if default dependencies added to main
self.main_program_default_deps = True
self.test_name = "" # name of testcase
self.expected_outputs = OrderedDict() # expected outputs of testcase
self.xfail = False # set if this is a XFAIL testcase
self.xtest = False # set if this is put under 'xtests'
# Add 'object -> [object, object, ...]' relations to CURR_MAP
def __add_deps_internal(self, src_objs, dst_objs, curr_map):
for src in src_objs:
for dst in dst_objs:
if not src in curr_map:
curr_map[src] = []
if not dst in curr_map[src]:
curr_map[src].append(dst)
def add_deps(self, src_objs, dst_objs):
self.__add_deps_internal(src_objs, dst_objs, self.deps)
def add_callrefs(self, src_objs, dst_objs):
self.__add_deps_internal(src_objs, dst_objs, self.callrefs)
# Process commands inside the {} construct.
# Note that throughout this script, the main program object is represented
# by the '#' string.
def process_main_program(test_descr, mainprog_str):
if mainprog_str:
test_descr.main_program = mainprog_str.split(';')
for s in test_descr.main_program:
m = re.match(r"^([+\-%^@])([0-9a-zA-Z]+)$", s)
if not m:
error("'%s' is not recognized main program operation" % (s))
opr = m.group(1)
obj = m.group(2)
if not obj in test_descr.objs:
test_descr.objs.append(obj)
if opr == '%' or opr == '@':
test_descr.add_callrefs(['#'], [obj])
# We have a main program specified, turn this off
test_descr.main_program_default_deps = False
# For(a1 a2 b1-12) object set descriptions, expand into an object list
def expand_object_set_string(descr_str):
obj_list = []
descr_list = descr_str.split()
for descr in descr_list:
m = re.match(r"^([a-zA-Z][0-9a-zA-Z]*)(-[0-9]+)?$", descr)
if not m:
error("'%s' is not a valid object set description" % (descr))
obj = m.group(1)
idx_end = m.group(2)
if not idx_end:
if not obj in obj_list:
obj_list.append(obj)
else:
idx_end = int(idx_end[1:])
m = re.match(r"^([0-9a-zA-Z][a-zA-Z]*)([0-9]+)$", obj)
if not m:
error("object description '%s' is malformed" % (obj))
obj_name = m.group(1)
idx_start = int(m.group (2))
if idx_start > idx_end:
error("index range %s-%s invalid" % (idx_start, idx_end))
for i in range(idx_start, idx_end + 1):
o = obj_name + str(i)
if not o in obj_list:
obj_list.append(o)
return obj_list
# Lexer for tokens
tokenspec = [ ("SONAME", r"soname\(([0-9a-zA-Z{}]+)\)=([0-9a-zA-Z]+)"),
("OBJ", r"([0-9a-zA-Z]+)"),
("DEP", r"->"),
("CALLREF", r"=>"),
("OBJSET", r"\[([0-9a-zA-Z]+)\]"),
("OBJSET2", r"\(([0-9a-zA-Z \-]+)\)"),
("OBJSET3", r"\*"),
("PROG", r"{([0-9a-zA-Z;+^\-%@]*)}"),
("PERMUTE", r"!"),
("SEMICOL", r";"),
("ERROR", r".") ]
tok_re = '|'.join('(?P<%s>%s)' % pair for pair in tokenspec)
# Main line parser of description language
def parse_description_string(t, descr_str):
# State used when parsing dependencies
curr_objs = []
in_dep = False
in_callref = False
def clear_dep_state():
nonlocal in_dep, in_callref
in_dep = in_callref = False
for m in re.finditer(tok_re, descr_str):
kind = m.lastgroup
value = m.group()
if kind == "SONAME":
s = re.match(r"soname\(([0-9a-zA-Z{}]+)\)=([0-9a-zA-Z]+)", value)
obj = s.group(1)
val = s.group(2)
if obj == "{}":
if '#' in t.soname_map:
error("soname of main program already set")
# Adjust to internal name
obj = '#'
else:
if re.match(r"[{}]", obj):
error("invalid object name '%s'" % (obj))
if not obj in t.objs:
error("'%s' is not name of already defined object" % (obj))
if obj in t.soname_map:
error("'%s' already has soname of '%s' set"
% (obj, t.soname_map[obj]))
t.soname_map[obj] = val
elif kind == "OBJ":
if in_dep:
t.add_deps(curr_objs, [value])
elif in_callref:
t.add_callrefs(curr_objs, [value])
clear_dep_state()
curr_objs = [value]
if not value in t.objs:
t.objs.append(value)
elif kind == "OBJSET":
objset = value[1:len(value)-1]
if in_dep:
t.add_deps(curr_objs, list (objset))
elif in_callref:
t.add_callrefs(curr_objs, list (objset))
clear_dep_state()
curr_objs = list(objset)
for o in list(objset):
if not o in t.objs:
t.objs.append(o)
elif kind == "OBJSET2":
descr_str = value[1:len(value)-1]
descr_str.strip()
objs = expand_object_set_string(descr_str)
if not objs:
error("empty object set '%s'" % (value))
if in_dep:
t.add_deps(curr_objs, objs)
elif in_callref:
t.add_callrefs(curr_objs, objs)
clear_dep_state()
curr_objs = objs
for o in objs:
if not o in t.objs:
t.objs.append(o)
elif kind == "OBJSET3":
if in_dep:
t.add_deps(curr_objs, ['*'])
elif in_callref:
t.add_callrefs(curr_objs, ['*'])
else:
error("non-dependence target set '*' can only be used "
"as target of ->/=> operations")
clear_dep_state()
curr_objs = ['*']
elif kind == "PERMUTE":
if in_dep or in_callref:
error("syntax error, permute operation invalid here")
if not curr_objs:
error("syntax error, no objects to permute here")
for obj in curr_objs:
if not obj in t.dep_permutations:
# Signal this object has permuted dependencies
t.dep_permutations[obj] = []
elif kind == "PROG":
if t.main_program:
error("cannot have more than one main program")
if in_dep:
error("objects cannot have dependency on main program")
if in_callref:
# TODO: A DSO can resolve to a symbol in the main binary,
# which we syntactically allow here, but haven't yet
# implemented.
t.add_callrefs(curr_objs, ["#"])
process_main_program(t, value[1:len(value)-1])
clear_dep_state()
curr_objs = ["#"]
elif kind == "DEP":
if in_dep or in_callref:
error("syntax error, multiple contiguous ->,=> operations")
if '*' in curr_objs:
error("non-dependence target set '*' can only be used "
"as target of ->/=> operations")
in_dep = True
elif kind == "CALLREF":
if in_dep or in_callref:
error("syntax error, multiple contiguous ->,=> operations")
if '*' in curr_objs:
error("non-dependence target set '*' can only be used "
"as target of ->/=> operations")
in_callref = True
elif kind == "SEMICOL":
curr_objs = []
clear_dep_state()
else:
error("unknown token '%s'" % (value))
return t
# Main routine to process each testcase description
def process_testcase(t):
global objpfx
assert t.test_name
base_test_name = t.test_name
test_subdir = base_test_name + "-dir"
testpfx = objpfx + test_subdir + "/"
test_srcdir = "dso-sort-tests-src/"
testpfx_src = objpfx + test_srcdir
if not os.path.exists(testpfx):
os.mkdir(testpfx)
if not os.path.exists(testpfx_src):
os.mkdir(testpfx_src)
def find_objs_not_depended_on(t):
objs_not_depended_on = []
for obj in t.objs:
skip = False
for r in t.deps.items():
if obj in r[1]:
skip = True
break
if not skip:
objs_not_depended_on.append(obj)
return objs_not_depended_on
non_dep_tgt_objs = find_objs_not_depended_on(t)
for obj in t.objs:
if obj in t.deps:
deps = t.deps[obj]
if '*' in deps:
deps.remove('*')
t.add_deps([obj], non_dep_tgt_objs)
if obj in t.callrefs:
deps = t.callrefs[obj]
if '*' in deps:
deps.remove('*')
t.add_callrefs([obj], non_dep_tgt_objs)
if "#" in t.deps:
deps = t.deps["#"]
if '*' in deps:
deps.remove('*')
t.add_deps(["#"], non_dep_tgt_objs)
# If no main program was specified in dependency description, make a
# default main program with deps pointing to all DSOs which are not
# depended by another DSO.
if t.main_program_default_deps:
main_deps = non_dep_tgt_objs
if not main_deps:
error("no objects for default main program to point "
"dependency to(all objects strongly connected?)")
t.add_deps(["#"], main_deps)
# Some debug output
if cmdlineargs.debug_output:
print("Testcase: %s" % (t.test_name))
print("All objects: %s" % (t.objs))
print("--- Static link dependencies ---")
for r in t.deps.items():
print("%s -> %s" % (r[0], r[1]))
print("--- Objects whose dependencies are to be permuted ---")
for r in t.dep_permutations.items():
print("%s" % (r[0]))
print("--- Call reference dependencies ---")
for r in t.callrefs.items():
print("%s => %s" % (r[0], r[1]))
print("--- main program ---")
print(t.main_program)
# Main testcase generation routine, does Makefile fragment generation,
# testcase source generation, and if --build specified builds testcase.
def generate_testcase(test_descr, test_suffix):
test_name = test_descr.test_name + test_suffix
# Print out needed Makefile fragments for use in glibc/elf/Makefile.
module_names = ""
for o in test_descr.objs:
rule = ("$(objpfx)" + test_subdir + "/" + test_name
+ "-" + o + ".os: $(objpfx)" + test_srcdir
+ test_name + "-" + o + ".c\n"
"\t$(compile.c) $(OUTPUT_OPTION)\n")
makefile.write (rule)
module_names += " " + test_subdir + "/" + test_name + "-" + o
makefile.write("modules-names +=%s\n" % (module_names))
# Depth-first traversal, executing FN(OBJ) in post-order
def dfs(t, fn):
def dfs_rec(obj, fn, obj_visited):
if obj in obj_visited:
return
obj_visited[obj] = True
if obj in t.deps:
for dep in t.deps[obj]:
dfs_rec(dep, fn, obj_visited)
fn(obj)
obj_visited = {}
for obj in t.objs:
dfs_rec(obj, fn, obj_visited)
# Generate link dependencies for all DSOs, done in a DFS fashion.
# Usually this doesn't need to be this complex, just listing the direct
# dependencies is enough. However to support creating circular
# dependency situations, traversing it by DFS and tracking processing
# status is the natural way to do it.
obj_processed = {}
fake_created = {}
def gen_link_deps(obj):
if obj in test_descr.deps:
dso = test_subdir + "/" + test_name + "-" + obj + ".so"
dependencies = ""
for dep in test_descr.deps[obj]:
if dep in obj_processed:
depstr = (" $(objpfx)" + test_subdir + "/"
+ test_name + "-" + dep + ".so")
else:
# A circular dependency is satisfied by making a
# fake DSO tagged with the correct SONAME
depstr = (" $(objpfx)" + test_subdir + "/"
+ test_name + "-" + dep + ".FAKE.so")
# Create empty C file and Makefile fragments for fake
# object. This only needs to be done at most once for
# an object name.
if not dep in fake_created:
f = open(testpfx_src + test_name + "-" + dep
+ ".FAKE.c", "w")
f.write(" \n")
f.close()
# Generate rule to create fake object
makefile.write \
("LDFLAGS-%s = -Wl,--no-as-needed "
"-Wl,-soname=%s\n"
% (test_name + "-" + dep + ".FAKE.so",
("$(objpfx)" + test_subdir + "/"
+ test_name + "-" + dep + ".so")))
rule = ("$(objpfx)" + test_subdir + "/"
+ test_name + "-" + dep + ".FAKE.os: "
"$(objpfx)" + test_srcdir
+ test_name + "-" + dep + ".FAKE.c\n"
"\t$(compile.c) $(OUTPUT_OPTION)\n")
makefile.write (rule)
makefile.write \
("modules-names += %s\n"
% (test_subdir + "/"
+ test_name + "-" + dep + ".FAKE"))
fake_created[dep] = True
dependencies += depstr
makefile.write("$(objpfx)%s:%s\n" % (dso, dependencies))
# Mark obj as processed
obj_processed[obj] = True
dfs(test_descr, gen_link_deps)
# Print LDFLAGS-* and *-no-z-defs
for o in test_descr.objs:
dso = test_name + "-" + o + ".so"
ldflags = "-Wl,--no-as-needed"
if o in test_descr.soname_map:
soname = ("$(objpfx)" + test_subdir + "/"
+ test_name + "-"
+ test_descr.soname_map[o] + ".so")
ldflags += (" -Wl,-soname=" + soname)
makefile.write("LDFLAGS-%s = %s\n" % (dso, ldflags))
if o in test_descr.callrefs:
makefile.write("%s-no-z-defs = yes\n" % (dso))
# Print dependencies for main test program.
depstr = ""
if '#' in test_descr.deps:
for o in test_descr.deps['#']:
depstr += (" $(objpfx)" + test_subdir + "/"
+ test_name + "-" + o + ".so")
makefile.write("$(objpfx)%s/%s:%s\n" % (test_subdir, test_name, depstr))
ldflags = "-Wl,--no-as-needed"
if '#' in test_descr.soname_map:
soname = ("$(objpfx)" + test_subdir + "/"
+ test_name + "-"
+ test_descr.soname_map['#'] + ".so")
ldflags += (" -Wl,-soname=" + soname)
makefile.write("LDFLAGS-%s = %s\n" % (test_name, ldflags))
rule = ("$(objpfx)" + test_subdir + "/" + test_name + ".o: "
"$(objpfx)" + test_srcdir + test_name + ".c\n"
"\t$(compile.c) $(OUTPUT_OPTION)\n")
makefile.write (rule)
# Ensure that all shared objects are built before running the
# test, whether there link-time dependencies or not.
depobjs = ["$(objpfx){}/{}-{}.so".format(test_subdir, test_name, dep)
for dep in test_descr.objs]
makefile.write("$(objpfx){}.out: {}\n".format(
base_test_name, " ".join(depobjs)))
# Add main executable to test-srcs
makefile.write("test-srcs += %s/%s\n" % (test_subdir, test_name))
# Add dependency on main executable of test
makefile.write("$(objpfx)%s.out: $(objpfx)%s/%s\n"
% (base_test_name, test_subdir, test_name))
for r in test_descr.expected_outputs.items():
tunable_options = []
specific_tunable = r[0]
xfail = r[1][1]
if specific_tunable != "":
tunable_options = [specific_tunable]
else:
tunable_options = default_tunable_options
if not tunable_options:
tunable_options = [""]
for tunable in tunable_options:
tunable_env = ""
tunable_sfx = ""
exp_tunable_sfx = ""
if tunable:
tunable_env = "GLIBC_TUNABLES=%s " % tunable
tunable_sfx = "-" + tunable.replace("=","_")
if specific_tunable:
tunable_sfx = "-" + specific_tunable.replace("=","_")
exp_tunable_sfx = tunable_sfx
tunable_descr = ("(%s)" % tunable_env.strip()
if tunable_env else "")
# Write out fragment of shell script for this single test.
test_descr.sh.write \
("%s${test_wrapper_env} ${run_program_env} \\\n"
"${common_objpfx}support/test-run-command \\\n"
"${common_objpfx}elf/ld.so \\\n"
"--library-path ${common_objpfx}elf/%s:"
"${common_objpfx}elf:${common_objpfx}.:"
"${common_objpfx}dlfcn \\\n"
"${common_objpfx}elf/%s/%s > \\\n"
" ${common_objpfx}elf/%s/%s%s.output\n"
% (tunable_env ,test_subdir,
test_subdir, test_name, test_subdir, test_name,
tunable_sfx))
# Generate a run of each test and compare with expected out
test_descr.sh.write \
("if [ $? -ne 0 ]; then\n"
" echo '%sFAIL: %s%s execution test'\n"
" something_failed=true\n"
"else\n"
" diff -wu ${common_objpfx}elf/%s/%s%s.output \\\n"
" ${common_objpfx}elf/%s%s%s.exp\n"
" if [ $? -ne 0 ]; then\n"
" echo '%sFAIL: %s%s expected output comparison'\n"
" something_failed=true\n"
" fi\n"
"fi\n"
% (("X" if xfail else ""), test_name, tunable_descr,
test_subdir, test_name, tunable_sfx,
test_srcdir, base_test_name, exp_tunable_sfx,
("X" if xfail else ""), test_name, tunable_descr))
# Generate C files according to dependency and calling relations from
# description string.
for obj in test_descr.objs:
src_name = test_name + "-" + obj + ".c"
f = open(testpfx_src + src_name, "w")
if obj in test_descr.callrefs:
called_objs = test_descr.callrefs[obj]
for callee in called_objs:
f.write("extern void fn_%s (void);\n" % (callee))
if len(obj) == 1:
f.write("extern int putchar(int);\n")
f.write("static void __attribute__((constructor)) " +
"init(void){putchar('%s');putchar('>');}\n" % (obj))
f.write("static void __attribute__((destructor)) " +
"fini(void){putchar('<');putchar('%s');}\n" % (obj))
else:
f.write('extern int printf(const char *, ...);\n')
f.write('static void __attribute__((constructor)) ' +
'init(void){printf("%s>");}\n' % (obj))
f.write('static void __attribute__((destructor)) ' +
'fini(void){printf("<%s");}\n' % (obj))
if obj in test_descr.callrefs:
called_objs = test_descr.callrefs[obj]
if len(obj) != 1:
f.write("extern int putchar(int);\n")
f.write("void fn_%s (void) {\n" % (obj))
if len(obj) == 1:
f.write(" putchar ('%s');\n" % (obj));
f.write(" putchar ('(');\n");
else:
f.write(' printf ("%s(");\n' % (obj));
for callee in called_objs:
f.write(" fn_%s ();\n" % (callee))
f.write(" putchar (')');\n");
f.write("}\n")
else:
for callref in test_descr.callrefs.items():
if obj in callref[1]:
if len(obj) == 1:
# We need to declare printf here in this case.
f.write('extern int printf(const char *, ...);\n')
f.write("void fn_%s (void) {\n" % (obj))
f.write(' printf ("%s()");\n' % (obj))
f.write("}\n")
break
f.close()
# Open C file for writing main program
f = open(testpfx_src + test_name + ".c", "w")
# if there are some operations in main(), it means we need -ldl
f.write("#include \n")
f.write("#include \n")
f.write("#include \n")
for s in test_descr.main_program:
if s[0] == '@':
f.write("extern void fn_%s (void);\n" % (s[1:]));
f.write("int main (void) {\n")
f.write(" putchar('{');\n")
# Helper routine for generating sanity checking code.
def put_fail_check(fail_cond, action_desc):
f.write(' if (%s) { printf ("\\n%s failed: %%s\\n", '
'dlerror()); exit (1);}\n' % (fail_cond, action_desc))
i = 0
while i < len(test_descr.main_program):
s = test_descr.main_program[i]
obj = s[1:]
dso = test_name + "-" + obj
if s[0] == '+' or s[0] == '^':
if s[0] == '+':
dlopen_flags = "RTLD_LAZY|RTLD_GLOBAL"
f.write(" putchar('+');\n");
else:
dlopen_flags = "RTLD_LAZY"
f.write(" putchar(':');\n");
if len(obj) == 1:
f.write(" putchar('%s');\n" % (obj));
else:
f.write(' printf("%s");\n' % (obj));
f.write(" putchar('[');\n");
f.write(' void *%s = dlopen ("%s.so", %s);\n'
% (obj, dso, dlopen_flags))
put_fail_check("!%s" % (obj),
"%s.so dlopen" % (dso))
f.write(" putchar(']');\n");
elif s[0] == '-':
f.write(" putchar('-');\n");
if len(obj) == 1:
f.write(" putchar('%s');\n" % (obj));
else:
f.write(' printf("%s");\n' % (obj));
f.write(" putchar('[');\n");
put_fail_check("dlclose (%s) != 0" % (obj),
"%s.so dlclose" % (dso))
f.write(" putchar(']');\n");
elif s[0] == '%':
f.write(" putchar('%');\n");
f.write(' void (*fn_%s)(void) = dlsym (%s, "fn_%s");\n'
% (obj, obj, obj))
put_fail_check("!fn_%s" % (obj),
"dlsym(fn_%s) from %s.so" % (obj, dso))
f.write(" fn_%s ();\n" % (obj))
elif s[0] == '@':
f.write(" putchar('@');\n");
f.write(" fn_%s ();\n" % (obj))
f.write(" putchar(';');\n");
i += 1
f.write(" putchar('}');\n")
f.write(" return 0;\n")
f.write("}\n")
f.close()
# --build option processing: build generated sources using 'build_gcc'
if cmdlineargs.build:
# Helper routine to run a shell command, for running GCC below
def run_cmd(args):
cmd = str.join(' ', args)
if cmdlineargs.debug_output:
print(cmd)
p = subprocess.Popen(args)
p.wait()
if p.returncode != 0:
error("error running command: %s" % (cmd))
# Compile individual .os files
for obj in test_descr.objs:
src_name = test_name + "-" + obj + ".c"
obj_name = test_name + "-" + obj + ".os"
run_cmd([build_gcc, "-c", "-fPIC", testpfx_src + src_name,
"-o", testpfx + obj_name])
obj_processed = {}
fake_created = {}
# Function to create -.so
def build_dso(obj):
obj_name = test_name + "-" + obj + ".os"
dso_name = test_name + "-" + obj + ".so"
deps = []
if obj in test_descr.deps:
for dep in test_descr.deps[obj]:
if dep in obj_processed:
deps.append(dep)
else:
deps.append(dep + ".FAKE")
if not dep in fake_created:
base_name = testpfx + test_name + "-" + dep
src_base_name = (testpfx_src + test_name
+ "-" + dep)
cmd = [build_gcc, "-Wl,--no-as-needed",
("-Wl,-soname=" + base_name + ".so"),
"-shared", base_name + ".FAKE.c",
"-o", src_base_name + ".FAKE.so"]
run_cmd(cmd)
fake_created[dep] = True
dso_deps = map(lambda d: testpfx + test_name + "-" + d + ".so",
deps)
cmd = [build_gcc, "-shared", "-o", testpfx + dso_name,
testpfx + obj_name, "-Wl,--no-as-needed"]
if obj in test_descr.soname_map:
soname = ("-Wl,-soname=" + testpfx + test_name + "-"
+ test_descr.soname_map[obj] + ".so")
cmd += [soname]
cmd += list(dso_deps)
run_cmd(cmd)
obj_processed[obj] = True
# Build all DSOs, this needs to be in topological dependency order,
# or link will fail
dfs(test_descr, build_dso)
# Build main program
deps = []
if '#' in test_descr.deps:
deps = test_descr.deps['#']
main_deps = map(lambda d: testpfx + test_name + "-" + d + ".so",
deps)
cmd = [build_gcc, "-Wl,--no-as-needed", "-o", testpfx + test_name,
testpfx_src + test_name + ".c", "-L%s" % (os.getcwd()),
"-Wl,-rpath-link=%s" % (os.getcwd())]
if '#' in test_descr.soname_map:
soname = ("-Wl,-soname=" + testpfx + test_name + "-"
+ test_descr.soname_map['#'] + ".so")
cmd += [soname]
cmd += list(main_deps)
run_cmd(cmd)
# Check if we need to enumerate permutations of dependencies
need_permutation_processing = False
if t.dep_permutations:
# Adjust dep_permutations into map of object -> dependency permutations
for r in t.dep_permutations.items():
obj = r[0]
if obj in t.deps and len(t.deps[obj]) > 1:
deps = t.deps[obj]
t.dep_permutations[obj] = list(itertools.permutations (deps))
need_permutation_processing = True
def enum_permutations(t, perm_list):
test_subindex = 1
curr_perms = []
def enum_permutations_rec(t, perm_list):
nonlocal test_subindex, curr_perms
if len(perm_list) >= 1:
curr = perm_list[0]
obj = curr[0]
perms = curr[1]
if not perms:
# This may be an empty list if no multiple dependencies to
# permute were found, skip to next in this case
enum_permutations_rec(t, perm_list[1:])
else:
for deps in perms:
t.deps[obj] = deps
permstr = "" if obj == "#" else obj + "_"
permstr += str.join('', deps)
curr_perms.append(permstr)
enum_permutations_rec(t, perm_list[1:])
curr_perms = curr_perms[0:len(curr_perms)-1]
else:
# t.deps is now instantiated with one dependency order
# permutation(across all objects that have multiple
# permutations), now process a testcase
generate_testcase(t, ("_" + str (test_subindex)
+ "-" + str.join('-', curr_perms)))
test_subindex += 1
enum_permutations_rec(t, perm_list)
# Create *.exp files with expected outputs
for r in t.expected_outputs.items():
sfx = ""
if r[0] != "":
sfx = "-" + r[0].replace("=","_")
f = open(testpfx_src + t.test_name + sfx + ".exp", "w")
(output, xfail) = r[1]
f.write('%s' % output)
f.close()
# Create header part of top-level testcase shell script, to wrap execution
# and output comparison together.
t.sh = open(testpfx_src + t.test_name + ".sh", "w")
t.sh.write("#!/bin/sh\n")
t.sh.write("# Test driver for %s, generated by "
"dso-ordering-test.py\n" % (t.test_name))
t.sh.write("common_objpfx=$1\n")
t.sh.write("test_wrapper_env=$2\n")
t.sh.write("run_program_env=$3\n")
t.sh.write("something_failed=false\n")
# Starting part of Makefile fragment
makefile.write("ifeq (yes,$(build-shared))\n")
if need_permutation_processing:
enum_permutations(t, list (t.dep_permutations.items()))
else:
# We have no permutations to enumerate, just process testcase normally
generate_testcase(t, "")
# If testcase is XFAIL, indicate so
if t.xfail:
makefile.write("test-xfail-%s = yes\n" % t.test_name)
# Output end part of Makefile fragment
expected_output_files = ""
for r in t.expected_outputs.items():
sfx = ""
if r[0] != "":
sfx = "-" + r[0].replace("=","_")
expected_output_files += " $(objpfx)%s%s%s.exp" % (test_srcdir,
t.test_name, sfx)
makefile.write \
("$(objpfx)%s.out: $(objpfx)%s%s.sh%s "
"$(common-objpfx)support/test-run-command\n"
% (t.test_name, test_srcdir, t.test_name,
expected_output_files))
makefile.write("\t$(SHELL) $< $(common-objpfx) '$(test-wrapper-env)' "
"'$(run-program-env)' > $@; $(evaluate-test)\n")
makefile.write("ifeq ($(run-built-tests),yes)\n")
if t.xtest:
makefile.write("xtests-special += $(objpfx)%s.out\n" % (t.test_name))
else:
makefile.write("tests-special += $(objpfx)%s.out\n" % (t.test_name))
makefile.write("endif\n")
makefile.write("endif\n")
# Write ending part of shell script generation
t.sh.write("if $something_failed; then\n"
" exit 1\n"
"else\n"
" echo '%sPASS: all tests for %s succeeded'\n"
" exit 0\n"
"fi\n" % (("X" if t.xfail else ""),
t.test_name))
t.sh.close()
# Decription file parsing
def parse_description_file(filename):
global default_tunable_options
global current_input_lineno
f = open(filename)
if not f:
error("cannot open description file %s" % (filename))
descrfile_lines = f.readlines()
t = None
for line in descrfile_lines:
p = re.compile(r"#.*$")
line = p.sub("", line) # Filter out comments
line = line.strip() # Remove excess whitespace
current_input_lineno += 1
m = re.match(r"^tunable_option:\s*(.*)$", line)
if m:
if m.group(1) == "":
error("tunable option cannot be empty")
default_tunable_options.append(m.group (1))
continue
m = re.match(r"^clear_tunables$", line)
if m:
default_tunable_options = []
continue
m = re.match(r"^([^:]+):\s*(.*)$", line)
if m:
lhs = m.group(1)
o = re.match(r"^output(.*)$", lhs)
xfail = False
if not o:
o = re.match(r"^xfail_output(.*)$", lhs)
if o:
xfail = True;
if o:
if not t:
error("output specification without testcase description")
tsstr = ""
if o.group(1):
ts = re.match(r"^\(([a-zA-Z0-9_.=]*)\)$", o.group (1))
if not ts:
error("tunable option malformed '%s'" % o.group(1))
tsstr = ts.group(1)
t.expected_outputs[tsstr] = (m.group(2), xfail)
# Any tunable option XFAILed means entire testcase
# is XFAIL/XPASS
t.xfail |= xfail
else:
if t:
# Starting a new test description, end and process
# current one.
process_testcase(t)
t = TestDescr()
x = re.match(r"^xtest\((.*)\)$", lhs)
if x:
t.xtest = True
t.test_name = x.group(1)
else:
t.test_name = lhs
descr_string = m.group(2)
parse_description_string(t, descr_string)
continue
else:
if line:
if not t:
error("no active testcase description")
parse_description_string(t, line)
# Process last completed test description
if t:
process_testcase(t)
# Setup Makefile output to file or stdout as selected
if output_makefile:
output_makefile_dir = os.path.dirname(output_makefile)
if output_makefile_dir:
os.makedirs(output_makefile_dir, exist_ok = True)
makefile = open(output_makefile, "w")
else:
makefile = open(sys.stdout.fileno (), "w")
# Finally, the main top-level calling of above parsing routines.
if description_file:
parse_description_file(description_file)
else:
t = TestDescr()
t.test_name = test_name
parse_description_string(t, description)
process_testcase(t)
# Close Makefile fragment output
makefile.close()