Tracking Failure Origins¶
The question of "Where does this value come from?" is fundamental for debugging. Which earlier variables could possibly have influenced the current erroneous state? And how did their values come to be?
When programmers read code during debugging, they scan it for potential origins of given values. This can be a tedious experience, notably, if the origins spread across multiple separate locations, possibly even in different modules. In this chapter, we thus investigate means to determine such origins automatically – by collecting data and control dependencies during program execution.
from bookutils import YouTubeVideo
YouTubeVideo("sjf3cOR0lcI")
Prerequisites
- You should have read the Introduction to Debugging.
- To understand how to compute dependencies automatically (the second half of this chapter), you will need
- advanced knowledge of Python semantics
- knowledge on how to instrument and transform code
- knowledge on how an interpreter works
import bookutils.setup
from bookutils import quiz, next_inputs, print_content
Synopsis¶
To use the code provided in this chapter, write
>>> from debuggingbook.Slicer import <identifier>
and then make use of the following features.
This chapter provides a Slicer
class to automatically determine and visualize dynamic flows and dependencies. When we say that a variable $x$ depends on a variable $y$ (and that $y$ flows into $x$), we distinguish two kinds of dependencies:
- Data dependency: $x$ is assigned a value computed from $y$.
- Control dependency: A statement involving $x$ is executed only because a condition involving $y$ was evaluated, influencing the execution path.
Such dependencies are crucial for debugging, as they allow determininh the origins of individual values (and notably incorrect values).
To determine dynamic dependencies in a function func
, use
with Slicer() as slicer:
<Some call to func()>
and then slicer.graph()
or slicer.code()
to examine dependencies.
You can also explicitly specify the functions to be instrumented, as in
with Slicer(func, func_1, func_2) as slicer:
<Some call to func()>
Here is an example. The demo()
function computes some number from x
:
>>> def demo(x: int) -> int:
>>> z = x
>>> while x <= z <= 64:
>>> z *= 2
>>> return z
By using with Slicer()
, we first instrument demo()
and then execute it:
>>> with Slicer() as slicer:
>>> demo(10)
After execution is complete, you can output slicer
to visualize the dependencies and flows as graph. Data dependencies are shown as black solid edges; control dependencies are shown as grey dashed edges. The arrows indicate influence: If $y$ depends on $x$ (and thus $x$ flows into $y$), then we have an arrow $x \rightarrow y$.
We see how the parameter x
flows into z
, which is returned after some computation that is control dependent on a <test>
involving z
.
>>> slicer
An alternate representation is slicer.code()
, annotating the instrumented source code with (backward) dependencies. Data dependencies are shown with <=
, control dependencies with <-
; locations (lines) are shown in parentheses.
>>> slicer.code()
* 1 def demo(x: int) -> int:
* 2 z = x # <= x (1)
* 3 while x <= z <= 64: # <= z (4), x (1), z (2)
* 4 z *= 2 # <= z (4), z (2); <- <test> (3)
* 5 return z # <= z (4)
Dependencies can also be retrieved programmatically. The dependencies()
method returns a Dependencies
object encapsulating the dependency graph.
The method all_vars()
returns all variables in the dependency graph. Each variable is encoded as a pair (name, location) where location is a pair (codename, lineno).
>>> slicer.dependencies().all_vars()
{('<demo() return value>', (<function __main__.demo(x: int) -> int>, 5)),
('<test>', (<function __main__.demo(x: int) -> int>, 3)),
('x', (<function __main__.demo(x: int) -> int>, 1)),
('z', (<function __main__.demo(x: int) -> int>, 2)),
('z', (<function __main__.demo(x: int) -> int>, 4))}
code()
and graph()
methods can also be applied on dependencies. The method backward_slice(var)
returns a backward slice for the given variable (again given as a pair (name, location)). To retrieve where z
in Line 2 came from, use:
>>> _, start_demo = inspect.getsourcelines(demo)
>>> start_demo
1
>>> slicer.dependencies().backward_slice(('z', (demo, start_demo + 1))).graph()
Here are the classes defined in this chapter. A Slicer
instruments a program, using a DependencyTracker
at run time to collect Dependencies
.
Dependencies¶
In the Introduction to debugging, we have seen how faults in a program state propagate to eventually become visible as failures. This induces a debugging strategy called tracking origins:
- We start with a single faulty state f – the failure.
- We determine f's origins – the parts of earlier states that could have caused the faulty state f.
- For each of these origins e, we determine whether they are faulty or not.
- For each of the faulty origins, we in turn determine their origins.
- If we find a part of the state that is faulty, yet has only correct origins, we have found the defect.
In all generality, a "part of the state" can be anything that can influence the program – some configuration setting, some database content, or the state of a device. Almost always, though, it is through individual variables that a part of the state manifests itself.
The good news is that variables do not take arbitrary values at arbitrary times – instead, they are set and accessed at precise moments in time, as determined by the program's semantics. This allows us to determine their origins by reading program code.
Let us assume you have a piece of code that reads as follows. The middle()
function is supposed to return the "middle" number of three values x
, y
, and z
– that is, the one number that neither is the minimum nor the maximum.
def middle(x, y, z):
if y < z:
if x < y:
return y
elif x < z:
return y
else:
if x > y:
return y
elif x > z:
return x
return z
In most cases, middle()
runs just fine:
m = middle(1, 2, 3)
m
In others, however, it returns the wrong value:
m = middle(2, 1, 3)
m
This is a typical debugging situation: You see a value that is erroneous; and you want to find out where it came from.
- In our case, we see that the erroneous value was returned from
middle()
, so we identify the fivereturn
statements inmiddle()
that the value could have come from. - The value returned is the value of
y
, and neitherx
,y
, norz
are altered during the execution ofmiddle()
. Hence, it must be one of the threereturn y
statements that is the origin ofm
. But which one?
For our small example, we can fire up an interactive debugger and simply step through the function; this reveals us the conditions evaluated and the return
statement executed.
import Debugger # minor dependency
with Debugger.Debugger():
middle(2, 1, 3)
We now see that it was the second return
statement that returned the incorrect value. But why was it executed after all? To this end, we can resort to the middle()
source code and have a look at those conditions that caused the return y
statement to be executed. Indeed, the conditions y < z
, x > y
, and finally x < z
again are origins of the returned value – and in turn have x
, y
, and z
as origins.
In our above reasoning about origins, we have encountered two kinds of origins:
- earlier data values (such as the value of
y
being returned) and - earlier control conditions (such as the
if
conditions governing thereturn y
statement).
The later parts of the state that can be influenced by such origins are said to be dependent on these origins. Speaking of variables, a variable $x$ depends on the value of a variable $y$ (written as $x \leftarrow y$) if a change in $y$ could affect the value of $x$.
We distinguish two kinds of dependencies $x \leftarrow y$, aligned with the two kinds of origins as outlined above:
- Data dependency: $x$ is assigned a value computed from $y$. In our example,
m
is data dependent on the return value ofmiddle()
. - Control dependency: A statement involving $x$ is executed only because a condition involving $y$ was evaluated, influencing the execution path. In our example, the value returned by
return y
is control dependent on the several conditions along its path, which involvex
,y
, andz
.
Let us examine these dependencies in more detail.
Visualizing Dependencies
Note: This is an excursion, diverting away from the main flow of the chapter. Unless you know what you are doing, you are encouraged to skip this part.
To illustrate our examples, we introduce a Dependencies
class that captures dependencies between variables at specific locations.
A Class for Dependencies¶
Dependencies
holds two dependency graphs. data
holds data dependencies, control
holds control dependencies.
Each of the two is organized as a dictionary holding nodes as keys and sets of nodes as values. Each node comes as a tuple
(variable_name, location)
where variable_name
is a string and location
is a pair
(func, lineno)
denoting a unique location in the code.
This is also reflected in the following type definitions:
Location = Tuple[Callable, int]
Node = Tuple[str, Location]
Dependency = Dict[Node, Set[Node]]
In this chapter, for many purposes, we need to lookup a function's location, source code, or simply definition. The class StackInspector
provides a number of convenience functions for this purpose.
from StackInspector import StackInspector
The Dependencies
class builds on StackInspector
to capture dependencies.
class Dependencies(StackInspector):
"""A dependency graph"""
def __init__(self,
data: Optional[Dependency] = None,
control: Optional[Dependency] = None) -> None:
"""
Create a dependency graph from `data` and `control`.
Both `data` and `control` are dictionaries
holding _nodes_ as keys and sets of nodes as values.
Each node comes as a tuple (variable_name, location)
where `variable_name` is a string
and `location` is a pair (function, lineno)
where `function` is a callable and `lineno` is a line number
denoting a unique location in the code.
"""
if data is None:
data = {}
if control is None:
control = {}
self.data = data
self.control = control
for var in self.data:
self.control.setdefault(var, set())
for var in self.control:
self.data.setdefault(var, set())
self.validate()
The validate()
method checks for consistency.
class Dependencies(Dependencies):
def validate(self) -> None:
"""Check dependency structure."""
assert isinstance(self.data, dict)
assert isinstance(self.control, dict)
for node in (self.data.keys()) | set(self.control.keys()):
var_name, location = node
assert isinstance(var_name, str)
func, lineno = location
assert callable(func)
assert isinstance(lineno, int)
The source()
method returns the source code for a given node.
test_deps = Dependencies()
test_deps.source(('z', (middle, 1)))
Drawing Dependencies¶
Both data and control form a graph between nodes, and can be visualized as such. We use the graphviz
package for creating such visualizations.
make_graph()
sets the basic graph attributes.
import html
class Dependencies(Dependencies):
NODE_COLOR = 'peachpuff'
FONT_NAME = 'Courier' # 'Fira Mono' may produce warnings in 'dot'
def make_graph(self,
name: str = "dependencies",
comment: str = "Dependencies") -> Digraph:
return Digraph(name=name, comment=comment,
graph_attr={
},
node_attr={
'style': 'filled',
'shape': 'box',
'fillcolor': self.NODE_COLOR,
'fontname': self.FONT_NAME
},
edge_attr={
'fontname': self.FONT_NAME
})
graph()
returns a graph visualization.
class Dependencies(Dependencies):
def graph(self, *, mode: str = 'flow') -> Digraph:
"""
Draw dependencies. `mode` is either
* `'flow'`: arrows indicate information flow (from A to B); or
* `'depend'`: arrows indicate dependencies (B depends on A)
"""
self.validate()
g = self.make_graph()
self.draw_dependencies(g, mode)
self.add_hierarchy(g)
return g
def _repr_mimebundle_(self, include: Any = None, exclude: Any = None) -> Any:
"""If the object is output in Jupyter, render dependencies as a SVG graph"""
return self.graph()._repr_mimebundle_(include, exclude)
The main part of graph drawing takes place in two methods, draw_dependencies()
and add_hierarchy()
.
draw_dependencies()
processes through the graph, adding nodes and edges from the dependencies.
class Dependencies(Dependencies):
def all_vars(self) -> Set[Node]:
"""Return a set of all variables (as `var_name`, `location`) in the dependencies"""
all_vars = set()
for var in self.data:
all_vars.add(var)
for source in self.data[var]:
all_vars.add(source)
for var in self.control:
all_vars.add(var)
for source in self.control[var]:
all_vars.add(source)
return all_vars
class Dependencies(Dependencies):
def draw_edge(self, g: Digraph, mode: str,
node_from: str, node_to: str, **kwargs: Any) -> None:
if mode == 'flow':
g.edge(node_from, node_to, **kwargs)
elif mode == 'depend':
g.edge(node_from, node_to, dir="back", **kwargs)
else:
raise ValueError("`mode` must be 'flow' or 'depend'")
def draw_dependencies(self, g: Digraph, mode: str) -> None:
for var in self.all_vars():
g.node(self.id(var),
label=self.label(var),
tooltip=self.tooltip(var))
if var in self.data:
for source in self.data[var]:
self.draw_edge(g, mode, self.id(source), self.id(var))
if var in self.control:
for source in self.control[var]:
self.draw_edge(g, mode, self.id(source), self.id(var),
style='dashed', color='grey')
draw_dependencies()
makes use of a few helper functions.
class Dependencies(Dependencies):
def id(self, var: Node) -> str:
"""Return a unique ID for `var`."""
id = ""
# Avoid non-identifier characters
for c in repr(var):
if c.isalnum() or c == '_':
id += c
if c == ':' or c == ',':
id += '_'
return id
def label(self, var: Node) -> str:
"""Render node `var` using HTML style."""
(name, location) = var
source = self.source(var)
title = html.escape(name)
if name.startswith('<'):
title = f'<I>{title}</I>'
label = f'<B>{title}</B>'
if source:
label += (f'<FONT POINT-SIZE="9.0"><BR/><BR/>'
f'{html.escape(source)}'
f'</FONT>')
label = f'<{label}>'
return label
def tooltip(self, var: Node) -> str:
"""Return a tooltip for node `var`."""
(name, location) = var
func, lineno = location
return f"{func.__name__}:{lineno}"
In the second part of graph drawing, add_hierarchy()
adds invisible edges to ensure that nodes with lower line numbers are drawn above nodes with higher line numbers.
class Dependencies(Dependencies):
def add_hierarchy(self, g: Digraph) -> Digraph:
"""Add invisible edges for a proper hierarchy."""
functions = self.all_functions()
for func in functions:
last_var = None
last_lineno = 0
for (lineno, var) in functions[func]:
if last_var is not None and lineno > last_lineno:
g.edge(self.id(last_var),
self.id(var),
style='invis')
last_var = var
last_lineno = lineno
return g
class Dependencies(Dependencies):
def all_functions(self) -> Dict[Callable, List[Tuple[int, Node]]]:
"""
Return mapping
{`function`: [(`lineno`, `var`), (`lineno`, `var`), ...], ...}
for all functions in the dependencies.
"""
functions: Dict[Callable, List[Tuple[int, Node]]] = {}
for var in self.all_vars():
(name, location) = var
func, lineno = location
if func not in functions:
functions[func] = []
functions[func].append((lineno, var))
for func in functions:
functions[func].sort()
return functions
Here comes the graph in all its glory:
def middle_deps() -> Dependencies:
return Dependencies({('z', (middle, 1)): set(), ('y', (middle, 1)): set(), ('x', (middle, 1)): set(), ('<test>', (middle, 2)): {('y', (middle, 1)), ('z', (middle, 1))}, ('<test>', (middle, 3)): {('y', (middle, 1)), ('x', (middle, 1))}, ('<test>', (middle, 5)): {('z', (middle, 1)), ('x', (middle, 1))}, ('<middle() return value>', (middle, 6)): {('y', (middle, 1))}}, {('z', (middle, 1)): set(), ('y', (middle, 1)): set(), ('x', (middle, 1)): set(), ('<test>', (middle, 2)): set(), ('<test>', (middle, 3)): {('<test>', (middle, 2))}, ('<test>', (middle, 5)): {('<test>', (middle, 3))}, ('<middle() return value>', (middle, 6)): {('<test>', (middle, 5))}})
middle_deps()
By default, the arrow direction indicates flows – an arrow A → B indicates that information from A flows into B (and thus the state in A causes the state in B). By setting the extra keyword parameter mode
to depend
instead of flow
(default), you can reverse these arrows; then an arrow A → B indicates A depends on B.
middle_deps().graph(mode='depend')
Slices¶
The method backward_slice(*critera, mode='cd')
returns a subset of dependencies, following dependencies backward from the given slicing criteria criteria
. These criteria can be
- variable names (such as
<test>
); or (function, lineno)
pairs (such as(middle, 3)
); or(var_name, (function, lineno))
(such as(
x, (middle, 1))
) locations.
The extra parameter mode
controls which dependencies are to be followed:
d
= data dependenciesc
= control dependencies
Criterion = Union[str, Location, Node]
class Dependencies(Dependencies):
def expand_criteria(self, criteria: List[Criterion]) -> List[Node]:
"""Return list of vars matched by `criteria`."""
all_vars = []
for criterion in criteria:
criterion_var = None
criterion_func = None
criterion_lineno = None
if isinstance(criterion, str):
criterion_var = criterion
elif len(criterion) == 2 and callable(criterion[0]):
criterion_func, criterion_lineno = criterion
elif len(criterion) == 2 and isinstance(criterion[0], str):
criterion_var = criterion[0]
criterion_func, criterion_lineno = criterion[1]
else:
raise ValueError("Invalid argument")
for var in self.all_vars():
(var_name, location) = var
func, lineno = location
name_matches = (criterion_func is None or
criterion_func == func or
criterion_func.__name__ == func.__name__)
location_matches = (criterion_lineno is None or
criterion_lineno == lineno)
var_matches = (criterion_var is None or
criterion_var == var_name)
if name_matches and location_matches and var_matches:
all_vars.append(var)
return all_vars
def backward_slice(self, *criteria: Criterion,
mode: str = 'cd', depth: int = -1) -> Dependencies:
"""
Create a backward slice from nodes `criteria`.
`mode` can contain 'c' (draw control dependencies)
and 'd' (draw data dependencies) (default: 'cd')
"""
data = {}
control = {}
queue = self.expand_criteria(criteria)
seen = set()
while len(queue) > 0 and depth != 0:
var = queue[0]
queue = queue[1:]
seen.add(var)
if 'd' in mode:
# Follow data dependencies
data[var] = self.data[var]
for next_var in data[var]:
if next_var not in seen:
queue.append(next_var)
else:
data[var] = set()
if 'c' in mode:
# Follow control dependencies
control[var] = self.control[var]
for next_var in control[var]:
if next_var not in seen:
queue.append(next_var)
else:
control[var] = set()
depth -= 1
return Dependencies(data, control)
Data Dependencies¶
Here is an example of a data dependency in our middle()
program. The value y
returned by middle()
comes from the value y
as originally passed as argument. We use arrows $x \leftarrow y$ to indicate that a variable $x$ depends on an earlier variable $y$:
Here, we can see that the value y
in the return statement is data dependent on the value of y
as passed to middle()
. An alternate interpretation of this graph is a data flow: The value of y
in the upper node flows into the value of y
in the lower node.
Since we consider the values of variables at specific locations in the program, such data dependencies can also be interpreted as dependencies between statements – the above return
statement thus is data dependent on the initialization of y
in the upper node.
Control Dependencies¶
Here is an example of a control dependency. The execution of the above return
statement is controlled by the earlier test x < z
. We use gray dashed lines to indicate control dependencies:
This test in turn is controlled by earlier tests, so the full chain of control dependencies looks like this:
Dependency Graphs¶
The above <test>
values (and their statements) are in turn also dependent on earlier data, namely the x
, y
, and z
values as originally passed. We can draw all data and control dependencies in a single graph, called a program dependency graph:
This graph now gives us an idea on how to proceed to track the origins of the middle()
return value at the bottom. Its value can come from any of the origins – namely the initialization of y
at the function call, or from the <test>
that controls it. This test in turn depends on x
and z
and their associated statements, which we now can check one after the other.
Note that all these dependencies in the graph are dynamic dependencies – that is, they refer to statements actually evaluated in the run at hand, as well as the decisions made in that very run. There also are static dependency graphs coming from static analysis of the code; but for debugging, dynamic dependencies specific to the failing run are more useful.
Showing Dependencies with Code¶
While a graph gives us a representation about which possible data and control flows to track, integrating dependencies with actual program code results in a compact representation that is easy to reason about.
Listing Dependencies
To show dependencies as text, we introduce a method format_var()
that shows a single node (a variable) as text. By default, a node is referenced as
NAME (FUNCTION:LINENO)
However, within a given function, it makes no sense to re-state the function name again and again, so we have a shorthand
NAME (LINENO)
to state a dependency to variable NAME
in line LINENO
.
class Dependencies(Dependencies):
def format_var(self, var: Node, current_func: Optional[Callable] = None) -> str:
"""Return string for `var` in `current_func`."""
name, location = var
func, lineno = location
if current_func and (func == current_func or func.__name__ == current_func.__name__):
return f"{name} ({lineno})"
else:
return f"{name} ({func.__name__}:{lineno})"
format_var()
is used extensively in the __str__()
string representation of dependencies, listing all nodes and their data (<=
) and control (<-
) dependencies.
class Dependencies(Dependencies):
def __str__(self) -> str:
"""Return string representation of dependencies"""
self.validate()
out = ""
for func in self.all_functions():
code_name = func.__name__
if out != "":
out += "\n"
out += f"{code_name}():\n"
all_vars = list(set(self.data.keys()) | set(self.control.keys()))
all_vars.sort(key=lambda var: var[1][1])
for var in all_vars:
(name, location) = var
var_func, var_lineno = location
var_code_name = var_func.__name__
if var_code_name != code_name:
continue
all_deps = ""
for (source, arrow) in [(self.data, "<="), (self.control, "<-")]:
deps = ""
for data_dep in source[var]:
if deps == "":
deps = f" {arrow} "
else:
deps += ", "
deps += self.format_var(data_dep, func)
if deps != "":
if all_deps != "":
all_deps += ";"
all_deps += deps
if all_deps == "":
continue
out += (" " +
self.format_var(var, func) +
all_deps + "\n")
return out
Here is a compact string representation of dependencies. We see how the (last) middle() return value
has a data dependency to y
in Line 1, and to the <test>
in Line 5.
print(middle_deps())
The __repr__()
method shows a raw form of dependencies, useful for creating dependencies from scratch.
class Dependencies(Dependencies):
def repr_var(self, var: Node) -> str:
name, location = var
func, lineno = location
return f"({repr(name)}, ({func.__name__}, {lineno}))"
def repr_deps(self, var_set: Set[Node]) -> str:
if len(var_set) == 0:
return "set()"
return ("{" +
", ".join(f"{self.repr_var(var)}"
for var in var_set) +
"}")
def repr_dependencies(self, vars: Dependency) -> str:
return ("{\n " +
",\n ".join(
f"{self.repr_var(var)}: {self.repr_deps(vars[var])}"
for var in vars) +
"}")
def __repr__(self) -> str:
"""Represent dependencies as a Python expression"""
# Useful for saving and restoring values
return (f"Dependencies(\n" +
f" data={self.repr_dependencies(self.data)},\n" +
f" control={self.repr_dependencies(self.control)})")
print(repr(middle_deps()))
An even more useful representation comes when integrating these dependencies as comments into the code. The method code(item_1, item_2, ...)
lists the given (function) items, including their dependencies; code()
lists all functions contained in the dependencies.
class Dependencies(Dependencies):
def code(self, *items: Callable, mode: str = 'cd') -> None:
"""
List `items` on standard output, including dependencies as comments.
If `items` is empty, all included functions are listed.
`mode` can contain 'c' (draw control dependencies) and 'd' (draw data dependencies)
(default: 'cd').
"""
if len(items) == 0:
items = cast(Tuple[Callable], self.all_functions().keys())
for i, item in enumerate(items):
if i > 0:
print()
self._code(item, mode)
def _code(self, item: Callable, mode: str) -> None:
# The functions in dependencies may be (instrumented) copies
# of the original function. Find the function with the same name.
func = item
for fn in self.all_functions():
if fn == item or fn.__name__ == item.__name__:
func = fn
break
all_vars = self.all_vars()
slice_locations = set(location for (name, location) in all_vars)
source_lines, first_lineno = inspect.getsourcelines(func)
n = first_lineno
for line in source_lines:
line_location = (func, n)
if line_location in slice_locations:
prefix = "* "
else:
prefix = " "
print(f"{prefix}{n:4} ", end="")
comment = ""
for (mode_control, source, arrow) in [
('d', self.data, '<='),
('c', self.control, '<-')
]:
if mode_control not in mode:
continue
deps = ""
for var in source:
name, location = var
if location == line_location:
for dep_var in source[var]:
if deps == "":
deps = arrow + " "
else:
deps += ", "
deps += self.format_var(dep_var, item)
if deps != "":
if comment != "":
comment += "; "
comment += deps
if comment != "":
line = line.rstrip() + " # " + comment
print_content(line.rstrip(), '.py')
print()
n += 1
The following listing shows such an integration. For each executed line (*
), we see its data (<=
) and control (<-
) dependencies, listing the associated variables and line numbers. The comment
# <= y (1); <- <test> (5)
for Line 6, for instance, states that the return value is data dependent on the value of y
in Line 1, and control dependent on the test in Line 5.
Again, one can easily follow these dependencies back to track where a value came from (data dependencies) and why a statement was executed (control dependency).
One important aspect of dependencies is that they not only point to specific sources and causes of failures – but that they also rule out parts of program and state as failures.
- In the above code, Lines 8 and later have no influence on the output, simply because they were not executed.
- Furthermore, we see that we can start our investigation with Line 6, because that is the last one executed.
- The data dependencies tell us that no statement has interfered with the value of
y
between the function call and its return. - Hence, the error must be in the conditions or the final
return
statement.
With this in mind, recall that our original invocation was middle(2, 1, 3)
. Why and how is the above code wrong?
Indeed, from the controlling conditions, we see that y < z
, x >= y
, and x < z
all hold. Hence, y <= x < z
holds, and it is x
, not y
, that should be returned.
Slices¶
Given a dependency graph for a particular variable, we can identify the subset of the program that could have influenced it – the so-called slice. In the above code listing, these code locations are highlighted with *
characters. Only these locations are part of the slice.
Slices are central to debugging for two reasons:
- First, they rule out those locations of the program that could not have an effect on the failure. Hence, these locations need not be investigated as it comes to searching for the defect. Nor do they need to be considered for a fix, as any change outside the program slice by construction cannot affect the failure.
- Second, they bring together possible origins that may be scattered across the code. Many dependencies in program code are non-local, with references to functions, classes, and modules defined in other locations, files, or libraries. A slice brings together all those locations in a single whole.
Here is an example of a slice – this time for our well-known remove_html_markup()
function from the introduction to debugging:
from Intro_Debugging import remove_html_markup
print_content(inspect.getsource(remove_html_markup), '.py')
When we invoke remove_html_markup()
as follows...
remove_html_markup('<foo>bar</foo>')
... we obtain the following dependencies:
Again, we can read such a graph forward (starting from, say, s
) or backward (starting from the return value). Starting forward, we see how the passed string s
flows into the for
loop, breaking s
into individual characters c
that are then checked on various occasions, before flowing into the out
return value. We also see how the various if
conditions are all influenced by c
, tag
, and quote
.
Which are the locations that set tag
to True? To this end, we compute the slice of tag
at tag = True
:
We see where the value of tag
comes from: from the characters c
in s
as well as quote
, which all cause it to be set. Again, we can combine these dependencies and the listing in a single, compact view. Note, again, that there are no other locations in the code that could possibly have affected tag
in our run.
Indeed, our dynamic slices reflect dependencies as they occurred within a single execution. As the execution changes, so do the dependencies.
Tracking Techniques¶
For the remainder of this chapter, let us investigate means to determine such dependencies automatically – by collecting them during program execution. The idea is that with a single Python call, we can collect the dependencies for some computation, and present them to programmers – as graphs or as code annotations, as shown above.
To track dependencies, for every variable, we need to keep track of its origins – where it obtained its value, and which tests controlled its assignments. There are two ways to do so:
- Wrapping Data Objects
- Wrapping Data Accesses
Wrapping Data Objects¶
One way to track origins is to wrap each value in a class that stores both a value and the origin of the value. If a variable x
is initialized to zero in Line 3, for instance, we could store it as
x = (value=0, origin=<Line 3>)
and if it is copied in, say, Line 5 to another variable y
, we could store this as
y = (value=0, origin=<Line 3, Line 5>)
Such a scheme would allow us to track origins and dependencies right within the variable.
In a language like Python, it is actually possibly to subclass from basic types. Here's how we create a MyInt
subclass of int
:
class MyInt(int):
def __new__(cls: Type, value: Any, *args: Any, **kwargs: Any) -> Any:
return super(cls, cls).__new__(cls, value)
def __repr__(self) -> str:
return f"{int(self)}"
n: MyInt = MyInt(5)
We can access n
just like any integer:
n, n + 1
However, we can also add extra attributes to it:
n.origin = "Line 5"
n.origin
Such a "wrapping" scheme has the advantage of leaving program code untouched – simply pass "wrapped" objects instead of the original values. However, it also has a number of drawbacks.
- First, we must make sure that the "wrapper" objects are still compatible with the original values – notably by converting them back whenever needed. (What happens if an internal Python function expects an
int
and gets aMyInt
instead?) - Second, we have to make sure that origins do not get lost during computations – which involves overloading operators such as
+
,-
,*
, and so on. (Right now,MyInt(1) + 1
gives us anint
object, not aMyInt
.) - Third, we have to do this for all data types of a language, which is pretty tedious.
- Fourth and last, however, we want to track whenever a value is assigned to another variable. Python has no support for this, and thus our dependencies will necessarily be incomplete.
Wrapping Data Accesses¶
An alternate way of tracking origins is to instrument the source code such that all data read and write operations are tracked. That is, the original data stays unchanged, but we change the code instead.
In essence, for every occurrence of a variable x
being read, we replace it with
_data.get('x', x) # returns x
and for every occurrence of a value being written to x
, we replace the value with
_data.set('x', value) # returns value
and let the _data
object track these reads and writes.
Hence, an assignment such as
a = b + c
would get rewritten to
a = _data.set('a', _data.get('b', b) + _data.get('c', c))
and with every access to _data
, we would track
- the current location in the code, and
- whether the respective variable was read or written.
For the above statement, we could deduce that b
and c
were read, and a
was written – which makes a
data dependent on b
and c
.
The advantage of such instrumentation is that it works with arbitrary objects (in Python, that is) – we do not care whether a
, b
, and c
would be integers, floats, strings, lists or any other type for which +
would be defined. Also, the code semantics remain entirely unchanged.
The disadvantage, however, is that it takes a bit of effort to exactly separate reads and writes into individual groups, and that a number of language features have to be handled separately. This is what we do in the remainder of this chapter.
A Data Tracker¶
To implement _data
accesses as shown above, we introduce the DataTracker
class. As its name suggests, it keeps track of variables being read and written, and provides methods to determine the code location where this took place.
class DataTracker(StackInspector):
"""Track data accesses during execution"""
def __init__(self, log: bool = False) -> None:
"""Constructor. If `log` is set, turn on logging."""
self.log = log
set()
is invoked when a variable is set, as in
pi = _data.set('pi', 3.1415)
By default, we simply log the access using name and value. (loads
will be used later.)
class DataTracker(DataTracker):
def set(self, name: str, value: Any, loads: Optional[Set[str]] = None) -> Any:
"""Track setting `name` to `value`."""
if self.log:
caller_func, lineno = self.caller_location()
print(f"{caller_func.__name__}:{lineno}: setting {name}")
return value
get()
is invoked when a variable is retrieved, as in
print(_data.get('pi', pi))
By default, we simply log the access.
class DataTracker(DataTracker):
def get(self, name: str, value: Any) -> Any:
"""Track getting `value` from `name`."""
if self.log:
caller_func, lineno = self.caller_location()
print(f"{caller_func.__name__}:{lineno}: getting {name}")
return value
Here's an example of a logging DataTracker
:
_test_data = DataTracker(log=True)
x = _test_data.set('x', 1)
_test_data.get('x', x)
Instrumenting Source Code¶
How do we transform source code such that read and write accesses to variables would be automatically rewritten? To this end, we inspect the internal representation of source code, namely the abstract syntax trees (ASTs). An AST represents the code as a tree, with specific node types for each syntactical element.
import ast
from bookutils import show_ast
Here is the tree representation for our middle()
function. It starts with a FunctionDef
node at the top (with the name "middle"
and the three arguments x
, y
, z
as children), followed by a subtree for each of the If
statements, each of which contains a branch for when their condition evaluates to True
and a branch for when their condition evaluates to False
.
middle_tree = ast.parse(inspect.getsource(middle))
show_ast(middle_tree)
At the very bottom of the tree, you can see a number of Name
nodes, referring individual variables. These are the ones we want to transform.
Tracking Variable Accesses¶
Our goal is to traverse the tree, identify all Name
nodes, and convert them to respective _data
accesses.
To this end, we manipulate the AST through the ast
Python module ast
. The official Python ast
reference is complete, but a bit brief; the documentation "Green Tree Snakes - the missing Python AST docs" provides an excellent introduction.
The Python ast
module provides a class NodeTransformer
that allows such transformations. By subclassing from it, we provide a method visit_Name()
that will be invoked for all Name
nodes – and replace it by a new subtree from make_get_data()
:
from ast import NodeTransformer, NodeVisitor, Name, AST
import typing
DATA_TRACKER = '_data'
def is_internal(id: str) -> bool:
"""Return True if `id` is a built-in function or type"""
return (id in dir(__builtins__) or id in dir(typing))
assert is_internal('int')
assert is_internal('None')
assert is_internal('Tuple')
class TrackGetTransformer(NodeTransformer):
def visit_Name(self, node: Name) -> AST:
self.generic_visit(node)
if is_internal(node.id):
# Do not change built-in names and types
return node
if node.id == DATA_TRACKER:
# Do not change own accesses
return node
if not isinstance(node.ctx, Load):
# Only change loads (not stores, not deletions)
return node
new_node = make_get_data(node.id)
ast.copy_location(new_node, node)
return new_node
Our function make_get_data(id, method)
returns a new subtree equivalent to the Python code _data.method('id', id)
.
from ast import Module, Load, Store, \
Attribute, With, withitem, keyword, Call, Expr, \
Assign, AugAssign, AnnAssign, Assert
# Starting with Python 3.8, these will become Constant.
# from ast import Num, Str, NameConstant
# Use `ast.Num`, `ast.Str`, and `ast.NameConstant` for compatibility
def make_get_data(id: str, method: str = 'get') -> Call:
return Call(func=Attribute(value=Name(id=DATA_TRACKER, ctx=Load()),
attr=method, ctx=Load()),
args=[ast.Str(id), Name(id=id, ctx=Load())],
keywords=[])
This is the tree that make_get_data()
produces:
show_ast(Module(body=[make_get_data("x")], type_ignores=[]))
How do we know that this is a correct subtree? We can carefully read the official Python ast
reference and then proceed by trial and error (and apply delta debugging to determine error causes). Or – pro tip! – we can simply take a piece of Python code, parse it and use ast.dump()
to print out how to construct the resulting AST:
print(ast.dump(ast.parse("_data.get('x', x)")))
If you compare the above output with the code of make_get_data()
, above, you will find out where the source of make_get_data()
comes from.
Let us put TrackGetTransformer
to action. Its visit()
method calls visit_Name()
, which then in turn transforms the Name
nodes as we want it. This happens in place.
TrackGetTransformer().visit(middle_tree);
To see the effect of our transformations, we introduce a method dump_tree()
which outputs the tree – and also compiles it to check for any inconsistencies.
def dump_tree(tree: AST) -> None:
print_content(ast.unparse(tree), '.py')
ast.fix_missing_locations(tree) # Must run this before compiling
_ = compile(cast(ast.Module, tree), '<dump_tree>', 'exec')
We see that our transformer has properly replaced all variable accesses:
dump_tree(middle_tree)
Let us now execute this code together with the DataTracker()
class we previously introduced. The class DataTrackerTester()
takes a (transformed) tree and a function. Using it as
with DataTrackerTester(tree, func):
func(...)
first executes the code in tree (possibly instrumenting func
) and then the with
body. At the end, func
is restored to its previous (non-instrumented) version.
from types import TracebackType
class DataTrackerTester:
def __init__(self, tree: AST, func: Callable, log: bool = True) -> None:
"""Constructor. Execute the code in `tree` while instrumenting `func`."""
# We pass the source file of `func` such that we can retrieve it
# when accessing the location of the new compiled code
source = cast(str, inspect.getsourcefile(func))
self.code = compile(cast(ast.Module, tree), source, 'exec')
self.func = func
self.log = log
def make_data_tracker(self) -> Any:
return DataTracker(log=self.log)
def __enter__(self) -> Any:
"""Rewrite function"""
tracker = self.make_data_tracker()
globals()[DATA_TRACKER] = tracker
exec(self.code, globals())
return tracker
def __exit__(self, exc_type: Type, exc_value: BaseException,
traceback: TracebackType) -> Optional[bool]:
"""Restore function"""
globals()[self.func.__name__] = self.func
del globals()[DATA_TRACKER]
return None
Here is our middle()
function:
print_content(inspect.getsource(middle), '.py', start_line_number=1)
And here is our instrumented middle_tree
executed with a DataTracker
object. We see how the middle()
tests access one argument after another.
with DataTrackerTester(middle_tree, middle):
middle(2, 1, 3)
After DataTrackerTester
is done, middle
is reverted to its non-instrumented version:
middle(2, 1, 3)
For a complete picture of what happens during executions, we implement a number of additional code transformers.
For each assignment statement x = y
, we change it to x = _data.set('x', y)
. This allows tracking assignments.
Tracking Assignments and Assertions
For the remaining transformers, we follow the same steps as for TrackGetTransformer
, except that our visit_...()
methods focus on different nodes, and return different subtrees. Here, we focus on assignment nodes.
We want to transform assignments x = value
into _data.set('x', value)
to track assignments to x
.
If the left-hand side of the assignment is more complex, as in x[y] = value
, we want to ensure the read access to x
and y
is also tracked. By transforming x[y] = value
into _data.set('x', value, loads=(x, y))
, we ensure that x
and y
are marked as read (as the otherwise ignored loads
argument would be changed to _data.get()
calls for x
and y
).
Using ast.dump()
, we reveal what the corresponding syntax tree has to look like:
print(ast.dump(ast.parse("_data.set('x', value, loads=(a, b))")))
Using this structure, we can write a function make_set_data()
which constructs such a subtree.
def make_set_data(id: str, value: Any,
loads: Optional[Set[str]] = None, method: str = 'set') -> Call:
"""
Construct a subtree _data.`method`('`id`', `value`).
If `loads` is set to [X1, X2, ...], make it
_data.`method`('`id`', `value`, loads=(X1, X2, ...))
"""
keywords=[]
if loads:
keywords = [
keyword(arg='loads',
value=ast.Tuple(
elts=[Name(id=load, ctx=Load()) for load in loads],
ctx=Load()
))
]
new_node = Call(func=Attribute(value=Name(id=DATA_TRACKER, ctx=Load()),
attr=method, ctx=Load()),
args=[ast.Str(id), value],
keywords=keywords)
ast.copy_location(new_node, value)
return new_node
The problem is, however: How do we get the name of the variable being assigned to? The left-hand side of an assignment can be a complex expression such as x[i]
. We use the leftmost name of the left-hand side as name to be assigned to.
class LeftmostNameVisitor(NodeVisitor):
def __init__(self) -> None:
super().__init__()
self.leftmost_name: Optional[str] = None
def visit_Name(self, node: Name) -> None:
if self.leftmost_name is None:
self.leftmost_name = node.id
self.generic_visit(node)
def leftmost_name(tree: AST) -> Optional[str]:
visitor = LeftmostNameVisitor()
visitor.visit(tree)
return visitor.leftmost_name
leftmost_name(ast.parse('a[x] = 25'))
Python also allows tuple assignments, as in (a, b, c) = (1, 2, 3)
. We extract all variables being stored (that is, expressions whose ctx
attribute is Store()
) and extract their (leftmost) names.
class StoreVisitor(NodeVisitor):
def __init__(self) -> None:
super().__init__()
self.names: Set[str] = set()
def visit(self, node: AST) -> None:
if hasattr(node, 'ctx') and isinstance(node.ctx, Store):
name = leftmost_name(node)
if name:
self.names.add(name)
self.generic_visit(node)
def store_names(tree: AST) -> Set[str]:
visitor = StoreVisitor()
visitor.visit(tree)
return visitor.names
store_names(ast.parse('a[x], b[y], c = 1, 2, 3'))
For complex assignments, we also want to access the names read in the left hand side of an expression.
class LoadVisitor(NodeVisitor):
def __init__(self) -> None:
super().__init__()
self.names: Set[str] = set()
def visit(self, node: AST) -> None:
if hasattr(node, 'ctx') and isinstance(node.ctx, Load):
name = leftmost_name(node)
if name is not None:
self.names.add(name)
self.generic_visit(node)
def load_names(tree: AST) -> Set[str]:
visitor = LoadVisitor()
visitor.visit(tree)
return visitor.names
load_names(ast.parse('a[x], b[y], c = 1, 2, 3'))
With this, we can now define TrackSetTransformer
as a transformer for regular assignments. Note that in Python, an assignment can have multiple targets, as in a = b = c
; we assign the data dependencies of c
to them all.
class TrackSetTransformer(NodeTransformer):
def visit_Assign(self, node: Assign) -> Assign:
value = ast.unparse(node.value)
if value.startswith(DATA_TRACKER + '.set'):
return node # Do not apply twice
for target in node.targets:
loads = load_names(target)
for store_name in store_names(target):
node.value = make_set_data(store_name, node.value,
loads=loads)
loads = set()
return node
The special form of "augmented assign" needs special treatment. We change statements of the form x += y
to x += _data.augment('x', y)
.
class TrackSetTransformer(TrackSetTransformer):
def visit_AugAssign(self, node: AugAssign) -> AugAssign:
value = ast.unparse(node.value)
if value.startswith(DATA_TRACKER):
return node # Do not apply twice
id = cast(str, leftmost_name(node.target))
node.value = make_set_data(id, node.value, method='augment')
return node
The corresponding augment()
method uses a combination of set()
and get()
to reflect the semantics.
class DataTracker(DataTracker):
def augment(self, name: str, value: Any) -> Any:
"""
Track augmenting `name` with `value`.
To be overloaded in subclasses.
"""
self.set(name, self.get(name, value))
return value
"Annotated" assignments come with a type, as in x: int = 3
. We treat them like regular assignments.
class TrackSetTransformer(TrackSetTransformer):
def visit_AnnAssign(self, node: AnnAssign) -> AnnAssign:
if node.value is None:
return node # just <var>: <type> without value
value = ast.unparse(node.value)
if value.startswith(DATA_TRACKER + '.set'):
return node # Do not apply twice
loads = load_names(node.target)
for store_name in store_names(node.target):
node.value = make_set_data(store_name, node.value,
loads=loads)
loads = set()
return node
Finally, we treat assertions just as if they were assignments to special variables:
class TrackSetTransformer(TrackSetTransformer):
def visit_Assert(self, node: Assert) -> Assert:
value = ast.unparse(node.test)
if value.startswith(DATA_TRACKER + '.set'):
return node # Do not apply twice
loads = load_names(node.test)
node.test = make_set_data("<assertion>", node.test, loads=loads)
return node
Here's all of these transformers in action. Our original function has a number of assignments:
def assign_test(x: int) -> Tuple[int, str]:
forty_two: int = 42
forty_two = forty_two = 42
a, b = 1, 2
c = d = [a, b]
c[d[a]].attr = 47
a *= b + 1
assert a > 0
return forty_two, "Forty-Two"
assign_tree = ast.parse(inspect.getsource(assign_test))
TrackSetTransformer().visit(assign_tree)
dump_tree(assign_tree)
If we later apply our transformer for data accesses, we can see that we track all variable reads and writes.
TrackGetTransformer().visit(assign_tree)
dump_tree(assign_tree)
Each return statement return x
is transformed to return _data.set('<return_value>', x)
. This allows tracking return values.
Tracking Return Values
Our TrackReturnTransformer
also makes use of make_set_data()
.
class TrackReturnTransformer(NodeTransformer):
def __init__(self) -> None:
self.function_name: Optional[str] = None
super().__init__()
def visit_FunctionDef(self, node: Union[ast.FunctionDef, ast.AsyncFunctionDef]) -> AST:
outer_name = self.function_name
self.function_name = node.name # Save current name
self.generic_visit(node)
self.function_name = outer_name
return node
def visit_AsyncFunctionDef(self, node: ast.AsyncFunctionDef) -> AST:
return self.visit_FunctionDef(node)
def return_value(self, tp: str = "return") -> str:
if self.function_name is None:
return f"<{tp} value>"
else:
return f"<{self.function_name}() {tp} value>"
def visit_return_or_yield(self, node: Union[ast.Return, ast.Yield, ast.YieldFrom],
tp: str = "return") -> AST:
if node.value is not None:
value = ast.unparse(node.value)
if not value.startswith(DATA_TRACKER + '.set'):
node.value = make_set_data(self.return_value(tp), node.value)
return node
def visit_Return(self, node: ast.Return) -> AST:
return self.visit_return_or_yield(node, tp="return")
def visit_Yield(self, node: ast.Yield) -> AST:
return self.visit_return_or_yield(node, tp="yield")
def visit_YieldFrom(self, node: ast.YieldFrom) -> AST:
return self.visit_return_or_yield(node, tp="yield")
This is the effect of TrackReturnTransformer
. We see that all return values are saved, and thus all locations of the corresponding return statements are tracked.
TrackReturnTransformer().visit(middle_tree)
dump_tree(middle_tree)
with DataTrackerTester(middle_tree, middle):
middle(2, 1, 3)
To track control dependencies, for every block controlled by an if
, while
, or for
:
- We wrap their tests in a
_data.test()
wrapper. This allows us to assign pseudo-variables like<test>
which hold the conditions. - We wrap their controlled blocks in a
with
statement. This allows us to track the variables read right before thewith
(= the controlling variables), and to restore the current controlling variables when the block is left.
A statement
if cond:
body
thus becomes
if _data.test(cond):
with _data:
body
Tracking Control
To modify control statements, we traverse the tree, looking for If
nodes:
class TrackControlTransformer(NodeTransformer):
def visit_If(self, node: ast.If) -> ast.If:
self.generic_visit(node)
node.test = self.make_test(node.test)
node.body = self.make_with(node.body)
node.orelse = self.make_with(node.orelse)
return node
The subtrees come from helper functions make_with()
and make_test()
. Again, all these subtrees are obtained via ast.dump()
.
class TrackControlTransformer(TrackControlTransformer):
def make_with(self, block: List[ast.stmt]) -> List[ast.stmt]:
"""Create a subtree 'with _data: `block`'"""
if len(block) == 0:
return []
block_as_text = ast.unparse(block[0])
if block_as_text.startswith('with ' + DATA_TRACKER):
return block # Do not apply twice
new_node = With(
items=[
withitem(
context_expr=Name(id=DATA_TRACKER, ctx=Load()),
optional_vars=None)
],
body=block
)
ast.copy_location(new_node, block[0])
return [new_node]
class TrackControlTransformer(TrackControlTransformer):
def make_test(self, test: ast.expr) -> ast.expr:
test_as_text = ast.unparse(test)
if test_as_text.startswith(DATA_TRACKER + '.test'):
return test # Do not apply twice
new_test = Call(func=Attribute(value=Name(id=DATA_TRACKER, ctx=Load()),
attr='test',
ctx=Load()),
args=[test],
keywords=[])
ast.copy_location(new_test, test)
return new_test
while
loops are handled just like if
constructs.
class TrackControlTransformer(TrackControlTransformer):
def visit_While(self, node: ast.While) -> ast.While:
self.generic_visit(node)
node.test = self.make_test(node.test)
node.body = self.make_with(node.body)
node.orelse = self.make_with(node.orelse)
return node
for
loops gets a different treatment, as there is no condition that would control the body. Still, we ensure that setting the iterator variable is properly tracked.
class TrackControlTransformer(TrackControlTransformer):
# regular `for` loop
def visit_For(self, node: Union[ast.For, ast.AsyncFor]) -> AST:
self.generic_visit(node)
id = ast.unparse(node.target).strip()
node.iter = make_set_data(id, node.iter)
# Uncomment if you want iterators to control their bodies
# node.body = self.make_with(node.body)
# node.orelse = self.make_with(node.orelse)
return node
# `for` loops in async functions
def visit_AsyncFor(self, node: ast.AsyncFor) -> AST:
return self.visit_For(node)
# `for` clause in comprehensions
def visit_comprehension(self, node: ast.comprehension) -> AST:
self.generic_visit(node)
id = ast.unparse(node.target).strip()
node.iter = make_set_data(id, node.iter)
return node
Here is the effect of TrackControlTransformer
:
TrackControlTransformer().visit(middle_tree)
dump_tree(middle_tree)
We extend DataTracker
to also log these events:
class DataTracker(DataTracker):
def test(self, cond: AST) -> AST:
"""Test condition `cond`. To be overloaded in subclasses."""
if self.log:
caller_func, lineno = self.caller_location()
print(f"{caller_func.__name__}:{lineno}: testing condition")
return cond
class DataTracker(DataTracker):
def __enter__(self) -> Any:
"""Enter `with` block. To be overloaded in subclasses."""
if self.log:
caller_func, lineno = self.caller_location()
print(f"{caller_func.__name__}:{lineno}: entering block")
return self
def __exit__(self, exc_type: Type, exc_value: BaseException,
traceback: TracebackType) -> Optional[bool]:
"""Exit `with` block. To be overloaded in subclasses."""
if self.log:
caller_func, lineno = self.caller_location()
print(f"{caller_func.__name__}:{lineno}: exiting block")
return None
with DataTrackerTester(middle_tree, middle):
middle(2, 1, 3)
We also want to be able to track calls across multiple functions. To this end, we wrap each call
func(arg1, arg2, ...)
into
_data.ret(_data.call(func)(_data.arg(arg1), _data.arg(arg2), ...))
each of which simply pass through their given argument, but which allow tracking the beginning of calls (call()
), the computation of arguments (arg()
), and the return of the call (ret()
), respectively.
Tracking Calls and Arguments
Our TrackCallTransformer
visits all Call
nodes, applying the transformations as shown above.
class TrackCallTransformer(NodeTransformer):
def make_call(self, node: AST, func: str,
pos: Optional[int] = None, kw: Optional[str] = None) -> Call:
"""Return _data.call(`func`)(`node`)"""
keywords = []
# `Num()` and `Str()` are deprecated in favor of `Constant()`
if pos:
keywords.append(keyword(arg='pos', value=ast.Num(pos)))
if kw:
keywords.append(keyword(arg='kw', value=ast.Str(kw)))
return Call(func=Attribute(value=Name(id=DATA_TRACKER,
ctx=Load()),
attr=func,
ctx=Load()),
args=[node],
keywords=keywords)
def visit_Call(self, node: Call) -> Call:
self.generic_visit(node)
call_as_text = ast.unparse(node)
if call_as_text.startswith(DATA_TRACKER + '.ret'):
return node # Already applied
func_as_text = ast.unparse(node)
if func_as_text.startswith(DATA_TRACKER + '.'):
return node # Own function
new_args = []
for n, arg in enumerate(node.args):
new_args.append(self.make_call(arg, 'arg', pos=n + 1))
node.args = cast(List[ast.expr], new_args)
for kw in node.keywords:
id = kw.arg if hasattr(kw, 'arg') else None
kw.value = self.make_call(kw.value, 'arg', kw=id)
node.func = self.make_call(node.func, 'call')
return self.make_call(node, 'ret')
Our example function middle()
does not contain any calls, but here is a function that invokes middle()
twice:
def test_call() -> int:
x = middle(1, 2, z=middle(1, 2, 3))
return x
call_tree = ast.parse(inspect.getsource(test_call))
dump_tree(call_tree)
If we invoke TrackCallTransformer
on this testing function, we get the following transformed code:
TrackCallTransformer().visit(call_tree);
dump_tree(call_tree)
def f() -> bool:
return math.isclose(1, 1.0)
f_tree = ast.parse(inspect.getsource(f))
dump_tree(f_tree)
TrackCallTransformer().visit(f_tree);
dump_tree(f_tree)
As before, our default arg()
, ret()
, and call()
methods simply log the event and pass through the given value.
class DataTracker(DataTracker):
def arg(self, value: Any, pos: Optional[int] = None, kw: Optional[str] = None) -> Any:
"""
Track `value` being passed as argument.
`pos` (if given) is the argument position (starting with 1).
`kw` (if given) is the argument keyword.
"""
if self.log:
caller_func, lineno = self.caller_location()
info = ""
if pos:
info += f" #{pos}"
if kw:
info += f" {repr(kw)}"
print(f"{caller_func.__name__}:{lineno}: pushing arg{info}")
return value
class DataTracker(DataTracker):
def ret(self, value: Any) -> Any:
"""Track `value` being used as return value."""
if self.log:
caller_func, lineno = self.caller_location()
print(f"{caller_func.__name__}:{lineno}: returned from call")
return value
class DataTracker(DataTracker):
def instrument_call(self, func: Callable) -> Callable:
"""Instrument a call to `func`. To be implemented in subclasses."""
return func
def call(self, func: Callable) -> Callable:
"""Track a call to `func`."""
if self.log:
caller_func, lineno = self.caller_location()
print(f"{caller_func.__name__}:{lineno}: calling {func}")
return self.instrument_call(func)
dump_tree(call_tree)
with DataTrackerTester(call_tree, test_call):
test_call()
test_call()
On the receiving end, for each function argument x
, we insert a call _data.param('x', x, [position info])
to initialize x
. This is useful for tracking parameters across function calls.
Tracking Parameters
Again, we use ast.dump()
to determine the correct syntax tree:
print(ast.dump(ast.parse("_data.param('x', x, pos=1, last=True)")))
class TrackParamsTransformer(NodeTransformer):
def visit_FunctionDef(self, node: ast.FunctionDef) -> ast.FunctionDef:
self.generic_visit(node)
named_args = []
for child in ast.iter_child_nodes(node.args):
if isinstance(child, ast.arg):
named_args.append(child)
create_stmts = []
for n, child in enumerate(named_args):
keywords=[keyword(arg='pos', value=ast.Num(n + 1))]
if child is node.args.vararg:
keywords.append(keyword(arg='vararg', value=ast.Str('*')))
if child is node.args.kwarg:
keywords.append(keyword(arg='vararg', value=ast.Str('**')))
if n == len(named_args) - 1:
keywords.append(keyword(arg='last',
value=ast.NameConstant(value=True)))
create_stmt = Expr(
value=Call(
func=Attribute(value=Name(id=DATA_TRACKER, ctx=Load()),
attr='param', ctx=Load()),
args=[ast.Str(child.arg),
Name(id=child.arg, ctx=Load())
],
keywords=keywords
)
)
ast.copy_location(create_stmt, node)
create_stmts.append(create_stmt)
node.body = cast(List[ast.stmt], create_stmts) + node.body
return node
This is the effect of TrackParamsTransformer()
. You see how the first three parameters are all initialized.
TrackParamsTransformer().visit(middle_tree)
dump_tree(middle_tree)
By default, the DataTracker
param()
method simply calls set()
to set variables.
class DataTracker(DataTracker):
def param(self, name: str, value: Any,
pos: Optional[int] = None, vararg: str = '', last: bool = False) -> Any:
"""
At the beginning of a function, track parameter `name` being set to `value`.
`pos` is the position of the argument (starting with 1).
`vararg` is "*" if `name` is a vararg parameter (as in *args),
and "**" is `name` is a kwargs parameter (as in *kwargs).
`last` is True if `name` is the last parameter.
"""
if self.log:
caller_func, lineno = self.caller_location()
info = ""
if pos is not None:
info += f" #{pos}"
print(f"{caller_func.__name__}:{lineno}: initializing {vararg}{name}{info}")
return self.set(name, value)
with DataTrackerTester(middle_tree, middle):
middle(2, 1, 3)
def args_test(x, *args, **kwargs):
print(x, *args, **kwargs)
args_tree = ast.parse(inspect.getsource(args_test))
TrackParamsTransformer().visit(args_tree)
dump_tree(args_tree)
with DataTrackerTester(args_tree, args_test):
args_test(1, 2, 3)
What do we obtain after we have applied all these transformers on middle()
? We see that the code now contains quite a load of instrumentation.
dump_tree(middle_tree)
And when we execute this code, we see that we can track quite a number of events, while the code semantics stay unchanged.
with DataTrackerTester(middle_tree, middle):
m = middle(2, 1, 3)
m
Transformer Stress Test
We stress test our transformers by instrumenting, transforming, and compiling a number of modules.
import Assertions # minor dependency
import Debugger # minor dependency
for module in [Assertions, Debugger, inspect, ast]:
module_tree = ast.parse(inspect.getsource(module))
assert isinstance(module_tree, ast.Module)
TrackCallTransformer().visit(module_tree)
TrackSetTransformer().visit(module_tree)
TrackGetTransformer().visit(module_tree)
TrackControlTransformer().visit(module_tree)
TrackReturnTransformer().visit(module_tree)
TrackParamsTransformer().visit(module_tree)
# dump_tree(module_tree)
ast.fix_missing_locations(module_tree) # Must run this before compiling
module_code = compile(module_tree, '<stress_test>', 'exec')
print(f"{repr(module.__name__)} instrumented successfully.")
Our next step will now be not only to log these events, but to actually construct dependencies from them.
Tracking Dependencies¶
To construct dependencies from variable accesses, we subclass DataTracker
into DependencyTracker
– a class that actually keeps track of all these dependencies. Its constructor initializes a number of variables we will discuss below.
class DependencyTracker(DataTracker):
"""Track dependencies during execution"""
def __init__(self, *args: Any, **kwargs: Any) -> None:
"""Constructor. Arguments are passed to DataTracker.__init__()"""
super().__init__(*args, **kwargs)
self.origins: Dict[str, Location] = {} # Where current variables were last set
self.data_dependencies: Dependency = {} # As with Dependencies, above
self.control_dependencies: Dependency = {}
self.last_read: List[str] = [] # List of last read variables
self.last_checked_location = (StackInspector.unknown, 1)
self._ignore_location_change = False
self.data: List[List[str]] = [[]] # Data stack
self.control: List[List[str]] = [[]] # Control stack
self.frames: List[Dict[Union[int, str], Any]] = [{}] # Argument stack
self.args: Dict[Union[int, str], Any] = {} # Current args
Data Dependencies¶
The first job of our DependencyTracker
is to construct dependencies between read and written variables.
Reading Variables¶
As in DataTracker
, the key method of DependencyTracker
again is get()
, invoked as _data.get('x', x)
whenever a variable x
is read. First and foremost, it appends the name of the read variable to the list last_read
.
class DependencyTracker(DependencyTracker):
def get(self, name: str, value: Any) -> Any:
"""Track a read access for variable `name` with value `value`"""
self.check_location()
self.last_read.append(name)
return super().get(name, value)
def check_location(self) -> None:
pass # More on that below
x = 5
y = 3
_test_data = DependencyTracker(log=True)
_test_data.get('x', x) + _test_data.get('y', y)
_test_data.last_read
Checking Locations¶
However, before appending the read variable to last_read
, _data.get()
does one more thing. By invoking check_location()
, it clears the last_read
list if we have reached a new line in the execution. This avoids situations such as
x
y
z = a + b
where x
and y
are, well, read, but do not affect the last line. Therefore, with every new line, the list of last read lines is cleared.
class DependencyTracker(DependencyTracker):
def clear_read(self) -> None:
"""Clear set of read variables"""
if self.log:
direct_caller = inspect.currentframe().f_back.f_code.co_name
caller_func, lineno = self.caller_location()
print(f"{caller_func.__name__}:{lineno}: "
f"clearing read variables {self.last_read} "
f"(from {direct_caller})")
self.last_read = []
def check_location(self) -> None:
"""If we are in a new location, clear set of read variables"""
location = self.caller_location()
func, lineno = location
last_func, last_lineno = self.last_checked_location
if self.last_checked_location != location:
if self._ignore_location_change:
self._ignore_location_change = False
elif func.__name__.startswith('<'):
# Entering list comprehension, eval(), exec(), ...
pass
elif last_func.__name__.startswith('<'):
# Exiting list comprehension, eval(), exec(), ...
pass
else:
# Standard case
self.clear_read()
self.last_checked_location = location
Two methods can suppress this reset of the last_read
list:
ignore_next_location_change()
suppresses the reset for the next line. This is useful when returning from a function, when the return value is still in the list of "read" variables.ignore_location_change()
suppresses the reset for the current line. This is useful if we already have returned from a function call.
class DependencyTracker(DependencyTracker):
def ignore_next_location_change(self) -> None:
self._ignore_location_change = True
def ignore_location_change(self) -> None:
self.last_checked_location = self.caller_location()
Watch how DependencyTracker
resets last_read
when a new line is executed:
_test_data = DependencyTracker()
_test_data.get('x', x) + _test_data.get('y', y)
_test_data.last_read
a = 42
b = -1
_test_data.get('a', a) + _test_data.get('b', b)
_test_data.last_read
Setting Variables¶
The method set()
creates dependencies. It is invoked as _data.set('x', value)
whenever a variable x
is set.
First and foremost, it takes the list of variables read last_read
, and for each of the variables $v$, it takes their origin $o$ (the place where they were last set) and appends the pair ($v$, $o$) to the list of data dependencies. It then does a similar thing with control dependencies (more on these below), and finally marks (in self.origins
) the current location of $v$.
import itertools
class DependencyTracker(DependencyTracker):
TEST = '<test>' # Name of pseudo-variables for testing conditions
def set(self, name: str, value: Any, loads: Optional[Set[str]] = None) -> Any:
"""Add a dependency for `name` = `value`"""
def add_dependencies(dependencies: Set[Node],
vars_read: List[str], tp: str) -> None:
"""Add origins of `vars_read` to `dependencies`."""
for var_read in vars_read:
if var_read in self.origins:
if var_read == self.TEST and tp == "data":
# Can't have data dependencies on conditions
continue
origin = self.origins[var_read]
dependencies.add((var_read, origin))
if self.log:
origin_func, origin_lineno = origin
caller_func, lineno = self.caller_location()
print(f"{caller_func.__name__}:{lineno}: "
f"new {tp} dependency: "
f"{name} <= {var_read} "
f"({origin_func.__name__}:{origin_lineno})")
self.check_location()
ret = super().set(name, value)
location = self.caller_location()
add_dependencies(self.data_dependencies.setdefault
((name, location), set()),
self.last_read, tp="data")
add_dependencies(self.control_dependencies.setdefault
((name, location), set()),
cast(List[str], itertools.chain.from_iterable(self.control)),
tp="control")
self.origins[name] = location
# Reset read info for next line
self.last_read = [name]
# Next line is a new location
self._ignore_location_change = False
return ret
def dependencies(self) -> Dependencies:
"""Return dependencies"""
return Dependencies(self.data_dependencies,
self.control_dependencies)
Let us illustrate set()
by example. Here's a set of variables read and written:
_test_data = DependencyTracker()
x = _test_data.set('x', 1)
y = _test_data.set('y', _test_data.get('x', x))
z = _test_data.set('z', _test_data.get('x', x) + _test_data.get('y', y))
The attribute origins
saves for each variable where it was last written:
_test_data.origins
The attribute data_dependencies
tracks for each variable the variables it was read from:
_test_data.data_dependencies
Hence, the above code already gives us a small dependency graph:
In the remainder of this section, we define methods to
- track control dependencies (
test()
,__enter__()
,__exit__()
) - track function calls and returns (
call()
,ret()
) - track function arguments (
arg()
,param()
) - check the validity of our dependencies (
validate()
).
Like our get()
and set()
methods above, these work by refining the appropriate methods defined in the DataTracker
class, building on our NodeTransformer
transformations.
Control Dependencies
Let us detail control dependencies. As discussed with DataTracker()
, we invoke test()
methods for all control conditions, and place the controlled blocks into with
clauses.
The test()
method simply sets a <test>
variable; this also places it in last_read
.
class DependencyTracker(DependencyTracker):
def test(self, value: Any) -> Any:
"""Track a test for condition `value`"""
self.set(self.TEST, value)
return super().test(value)
When entering a with
block, the set of last_read
variables holds the <test>
variable read. We save it in the control
stack, with the effect of any further variables written now being marked as controlled by <test>
.
class DependencyTracker(DependencyTracker):
def __enter__(self) -> Any:
"""Track entering an if/while/for block"""
self.control.append(self.last_read)
self.clear_read()
return super().__enter__()
When we exit the with
block, we restore earlier last_read
values, preparing for else
blocks.
class DependencyTracker(DependencyTracker):
def __exit__(self, exc_type: Type, exc_value: BaseException,
traceback: TracebackType) -> Optional[bool]:
"""Track exiting an if/while/for block"""
self.clear_read()
self.last_read = self.control.pop()
self.ignore_next_location_change()
return super().__exit__(exc_type, exc_value, traceback)
Here's an example of all these parts in action:
_test_data = DependencyTracker()
x = _test_data.set('x', 1)
y = _test_data.set('y', _test_data.get('x', x))
if _test_data.test(_test_data.get('x', x) >= _test_data.get('y', y)):
with _test_data:
z = _test_data.set('z',
_test_data.get('x', x) + _test_data.get('y', y))
_test_data.control_dependencies
The control dependency for z
is reflected in the dependency graph:
Calls and Returns
import copy
To handle complex expressions involving functions, we introduce a data stack. Every time we invoke a function func
(call()
is invoked), we save the list of current variables read last_read
on the data
stack; when we return (ret()
is invoked), we restore last_read
. This also ensures that only those variables read while evaluating arguments will flow into the function call.
class DependencyTracker(DependencyTracker):
def call(self, func: Callable) -> Callable:
"""Track a call of function `func`"""
func = super().call(func)
if inspect.isgeneratorfunction(func):
return self.call_generator(func)
# Save context
if self.log:
caller_func, lineno = self.caller_location()
print(f"{caller_func.__name__}:{lineno}: "
f"saving read variables {self.last_read}")
self.data.append(self.last_read)
self.clear_read()
self.ignore_next_location_change()
self.frames.append(self.args)
self.args = {}
return func
class DependencyTracker(DependencyTracker):
def ret(self, value: Any) -> Any:
"""Track a function return"""
value = super().ret(value)
if self.in_generator():
return self.ret_generator(value)
# Restore old context and add return value
ret_name = None
for var in self.last_read:
if var.startswith("<"): # "<return value>"
ret_name = var
self.last_read = self.data.pop()
if ret_name:
self.last_read.append(ret_name)
if self.args:
# We return from an uninstrumented function:
# Make return value depend on all args
for key, deps in self.args.items():
self.last_read += deps
self.ignore_location_change()
self.args = self.frames.pop()
if self.log:
caller_func, lineno = self.caller_location()
print(f"{caller_func.__name__}:{lineno}: "
f"restored read variables {self.last_read}")
return value
Generator functions (those which yield
a value) are not "called" in the sense that Python transfers control to them; instead, a "call" to a generator function creates a generator that is evaluated on demand. We mark generator function "calls" by saving None
on the stacks. When the generator function returns the generator, we wrap the generator such that the arguments are being restored when it is invoked.
class DependencyTracker(DependencyTracker):
def in_generator(self) -> bool:
"""True if we are calling a generator function"""
return len(self.data) > 0 and self.data[-1] is None
def call_generator(self, func: Callable) -> Callable:
"""Track a call of a generator function"""
# Mark the fact that we're in a generator with `None` values
self.data.append(None)
self.frames.append(None)
assert self.in_generator()
self.clear_read()
return func
def ret_generator(self, generator: Any) -> Any:
"""Track the return of a generator function"""
# Pop the two 'None' values pushed earlier
self.data.pop()
self.frames.pop()
if self.log:
caller_func, lineno = self.caller_location()
print(f"{caller_func.__name__}:{lineno}: "
f"wrapping generator {generator} (args={self.args})")
# At this point, we already have collected the args.
# The returned generator depends on all of them.
for arg in self.args:
self.last_read += self.args[arg]
# Wrap the generator such that the args are restored
# when it is actually invoked, such that we can map them
# to parameters.
saved_args = copy.deepcopy(self.args)
def wrapper() -> Generator[Any, None, None]:
self.args = saved_args
if self.log:
caller_func, lineno = self.caller_location()
print(f"{caller_func.__name__}:{lineno}: "
f"calling generator (args={self.args})")
self.ignore_next_location_change()
yield from generator
return wrapper()
We see an example of how function calls and returns work in conjunction with function arguments, discussed in the next section.
Function Arguments
Finally, we handle parameters and arguments. The args
stack holds the current stack of function arguments, holding the last_read
variable for each argument.
class DependencyTracker(DependencyTracker):
def arg(self, value: Any, pos: Optional[int] = None, kw: Optional[str] = None) -> Any:
"""
Track passing an argument `value`
(with given position `pos` 1..n or keyword `kw`)
"""
if self.log:
caller_func, lineno = self.caller_location()
print(f"{caller_func.__name__}:{lineno}: "
f"saving args read {self.last_read}")
if pos:
self.args[pos] = self.last_read
if kw:
self.args[kw] = self.last_read
self.clear_read()
return super().arg(value, pos, kw)
When accessing the arguments (with param()
), we can retrieve this set of read variables for each argument.
class DependencyTracker(DependencyTracker):
def param(self, name: str, value: Any,
pos: Optional[int] = None, vararg: str = "", last: bool = False) -> Any:
"""
Track getting a parameter `name` with value `value`
(with given position `pos`).
vararg parameters are indicated by setting `varargs` to
'*' (*args) or '**' (**kwargs)
"""
self.clear_read()
if vararg == '*':
# We over-approximate by setting `args` to _all_ positional args
for index in self.args:
if isinstance(index, int) and pos is not None and index >= pos:
self.last_read += self.args[index]
elif vararg == '**':
# We over-approximate by setting `kwargs` to _all_ passed keyword args
for index in self.args:
if isinstance(index, str):
self.last_read += self.args[index]
elif name in self.args:
self.last_read = self.args[name]
elif pos in self.args:
self.last_read = self.args[pos]
if self.log:
caller_func, lineno = self.caller_location()
print(f"{caller_func.__name__}:{lineno}: "
f"restored params read {self.last_read}")
self.ignore_location_change()
ret = super().param(name, value, pos)
if last:
self.clear_read()
self.args = {} # Mark `args` as processed
return ret
Let us illustrate all these on a small example.
def call_test() -> int:
c = 47
def sq(n: int) -> int:
return n * n
def gen(e: int) -> Generator[int, None, None]:
yield e * c
def just_x(x: Any, y: Any) -> Any:
return x
a = 42
b = gen(a)
d = list(b)[0]
xs = [1, 2, 3, 4]
ys = [sq(elem) for elem in xs if elem > 2]
return just_x(just_x(d, y=b), ys[0])
call_test()
We apply all our transformers on this code:
call_tree = ast.parse(inspect.getsource(call_test))
TrackCallTransformer().visit(call_tree)
TrackSetTransformer().visit(call_tree)
TrackGetTransformer().visit(call_tree)
TrackControlTransformer().visit(call_tree)
TrackReturnTransformer().visit(call_tree)
TrackParamsTransformer().visit(call_tree)
dump_tree(call_tree)
Again, we capture the dependencies:
class DependencyTrackerTester(DataTrackerTester):
def make_data_tracker(self) -> DependencyTracker:
return DependencyTracker(log=self.log)
with DependencyTrackerTester(call_tree, call_test, log=False) as call_deps:
call_test()
We see how
a
flows into the generatorb
and into the parametere
ofgen()
.xs
flows intoelem
which in turn flows into the parametern
ofsq()
. Both flow intoys
.just_x()
returns only itsx
argument.
call_deps.dependencies()
The code()
view lists each function separately:
call_deps.dependencies().code()
Diagnostics
To check the dependencies we obtain, we perform some minimal checks on whether a referenced variable actually also occurs in the source code.
import re
validate()
is automatically called whenever dependencies are output, so if you see any of its error messages, something may be wrong.
At this point, DependencyTracker
is complete; we have all in place to track even complex dependencies in instrumented code.
Slicing Code¶
Let us now put all these pieces together. We have a means to instrument the source code (our various NodeTransformer
classes) and a means to track dependencies (the DependencyTracker
class). Now comes the time to put all these things together in a single tool, which we call Slicer
.
The basic idea of Slicer
is that you can use it as follows:
with Slicer(func_1, func_2, ...) as slicer:
func(...)
which first instruments the functions given in the constructor (i.e., replaces their definitions with instrumented counterparts), and then runs the code in the body, calling instrumented functions, and allowing the slicer to collect dependencies. When the body returns, the original definition of the instrumented functions is restored.
An Instrumenter Base Class¶
The basic functionality of instrumenting a number of functions (and restoring them at the end of the with
block) comes in a Instrumenter
base class. It invokes instrument()
on all items to instrument; this is to be overloaded in subclasses.
class Instrumenter(StackInspector):
"""Instrument functions for dynamic tracking"""
def __init__(self, *items_to_instrument: Callable,
globals: Optional[Dict[str, Any]] = None,
log: Union[bool, int] = False) -> None:
"""
Create an instrumenter.
`items_to_instrument` is a list of items to instrument.
`globals` is a namespace to use (default: caller's globals())
"""
self.log = log
self.items_to_instrument: List[Callable] = list(items_to_instrument)
self.instrumented_items: Set[Any] = set()
if globals is None:
globals = self.caller_globals()
self.globals = globals
def __enter__(self) -> Any:
"""Instrument sources"""
items = self.items_to_instrument
if not items:
items = self.default_items_to_instrument()
for item in items:
self.instrument(item)
return self
def default_items_to_instrument(self) -> List[Callable]:
return []
def instrument(self, item: Any) -> Any:
"""Instrument `item`. To be overloaded in subclasses."""
if self.log:
print("Instrumenting", item)
self.instrumented_items.add(item)
return item
At the end of the with
block, we restore the given functions.
class Instrumenter(Instrumenter):
def __exit__(self, exc_type: Type, exc_value: BaseException,
traceback: TracebackType) -> Optional[bool]:
"""Restore sources"""
self.restore()
return None
def restore(self) -> None:
for item in self.instrumented_items:
self.globals[item.__name__] = item
By default, an Instrumenter
simply outputs a log message:
with Instrumenter(middle, log=True) as ins:
pass
The Slicer Class¶
The Slicer
class comes as a subclass of Instrumenter
. It sets its own dependency tracker (which can be overwritten by setting the dependency_tracker
keyword argument).
class Slicer(Instrumenter):
"""Track dependencies in an execution"""
def __init__(self, *items_to_instrument: Any,
dependency_tracker: Optional[DependencyTracker] = None,
globals: Optional[Dict[str, Any]] = None,
log: Union[bool, int] = False):
"""Create a slicer.
`items_to_instrument` are Python functions or modules with source code.
`dependency_tracker` is the tracker to be used (default: DependencyTracker).
`globals` is the namespace to be used(default: caller's `globals()`)
`log`=True or `log` > 0 turns on logging
"""
super().__init__(*items_to_instrument, globals=globals, log=log)
if dependency_tracker is None:
dependency_tracker = DependencyTracker(log=(log > 1))
self.dependency_tracker = dependency_tracker
self.saved_dependencies = None
def default_items_to_instrument(self) -> List[Callable]:
raise ValueError("Need one or more items to instrument")
The parse()
method parses a given item, returning its AST.
class Slicer(Slicer):
def parse(self, item: Any) -> AST:
"""Parse `item`, returning its AST"""
source_lines, lineno = inspect.getsourcelines(item)
source = "".join(source_lines)
if self.log >= 2:
print_content(source, '.py', start_line_number=lineno)
print()
print()
tree = ast.parse(source)
ast.increment_lineno(tree, lineno - 1)
return tree
The transform()
method applies the list of transformers defined earlier in this chapter.
class Slicer(Slicer):
def transformers(self) -> List[NodeTransformer]:
"""List of transformers to apply. To be extended in subclasses."""
return [
TrackCallTransformer(),
TrackSetTransformer(),
TrackGetTransformer(),
TrackControlTransformer(),
TrackReturnTransformer(),
TrackParamsTransformer()
]
def transform(self, tree: AST) -> AST:
"""Apply transformers on `tree`. May be extended in subclasses."""
# Apply transformers
for transformer in self.transformers():
if self.log >= 3:
print(transformer.__class__.__name__ + ':')
transformer.visit(tree)
ast.fix_missing_locations(tree)
if self.log >= 3:
print_content(ast.unparse(tree), '.py')
print()
print()
if 0 < self.log < 3:
print_content(ast.unparse(tree), '.py')
print()
print()
return tree
The execute()
method executes the transformed tree (such that we get the new definitions). We also make the dependency tracker available for the code in the with
block.
class Slicer(Slicer):
def execute(self, tree: AST, item: Any) -> None:
"""Compile and execute `tree`. May be extended in subclasses."""
# We pass the source file of `item` such that we can retrieve it
# when accessing the location of the new compiled code
source = cast(str, inspect.getsourcefile(item))
code = compile(cast(ast.Module, tree), source, 'exec')
# Enable dependency tracker
self.globals[DATA_TRACKER] = self.dependency_tracker
# Execute the code, resulting in a redefinition of item
exec(code, self.globals)
The instrument()
method puts all these together, first parsing the item into a tree, then transforming and executing the tree.
class Slicer(Slicer):
def instrument(self, item: Any) -> Any:
"""Instrument `item`, transforming its source code, and re-defining it."""
if is_internal(item.__name__):
return item # Do not instrument `print()` and the like
if inspect.isbuiltin(item):
return item # No source code
item = super().instrument(item)
tree = self.parse(item)
tree = self.transform(tree)
self.execute(tree, item)
new_item = self.globals[item.__name__]
return new_item
When we restore the original definition (after the with
block), we save the dependency tracker again.
class Slicer(Slicer):
def restore(self) -> None:
"""Restore original code."""
if DATA_TRACKER in self.globals:
self.saved_dependencies = self.globals[DATA_TRACKER]
del self.globals[DATA_TRACKER]
super().restore()
Three convenience functions allow us to see the dependencies as (well) dependencies, as code, and as graph. These simply invoke the respective functions on the saved dependencies.
class Slicer(Slicer):
def dependencies(self) -> Dependencies:
"""Return collected dependencies."""
if self.saved_dependencies is None:
return Dependencies({}, {})
return self.saved_dependencies.dependencies()
def code(self, *args: Any, **kwargs: Any) -> None:
"""Show code of instrumented items, annotated with dependencies."""
first = True
for item in self.instrumented_items:
if not first:
print()
self.dependencies().code(item, *args, **kwargs)
first = False
def graph(self, *args: Any, **kwargs: Any) -> Digraph:
"""Show dependency graph."""
return self.dependencies().graph(*args, **kwargs)
def _repr_mimebundle_(self, include: Any = None, exclude: Any = None) -> Any:
"""If the object is output in Jupyter, render dependencies as a SVG graph"""
return self.graph()._repr_mimebundle_(include, exclude)
Let us put Slicer
into action. We track our middle()
function:
with Slicer(middle) as slicer:
m = middle(2, 1, 3)
m
These are the dependencies in string form (used when printed):
print(slicer.dependencies())
This is the code form:
slicer.code()
And this is the graph form:
slicer
You can also access the raw repr()
form, which allows you to reconstruct dependencies at any time. (This is how we showed off dependencies at the beginning of this chapter, before even introducing the code that computes them.)
print(repr(slicer.dependencies()))
Diagnostics¶
The Slicer
constructor accepts a log
argument (default: False), which can be set to show various intermediate results:
log=True
(orlog=1
): Show instrumented source codelog=2
: Also log executionlog=3
: Also log individual transformer stepslog=4
: Also log source line numbers
More Examples¶
Let us demonstrate our Slicer
class on a few more examples.
Square Root¶
The square_root()
function from the chapter on assertions demonstrates a nice interplay between data and control dependencies.
import math
from Assertions import square_root # minor dependency
Here is the original source code:
print_content(inspect.getsource(square_root), '.py')
Turning on logging shows the instrumented version:
with Slicer(square_root, log=True) as root_slicer:
y = square_root(2.0)
The dependency graph shows how guess
and approx
flow into each other until they are the same.
root_slicer
Again, we can show the code annotated with dependencies:
root_slicer.code()
The astute reader may find that a statement assert p
does not control the following code, although it would be equivalent to if not p: raise Exception
. Why is that?
Indeed: we treat assertions as "neutral" in the sense that they do not affect the remainder of the program – if they are turned off, they have no effect; and if they are turned on, the remaining program logic should not depend on them. (Our instrumentation also has no special treatment of raise
or even return
statements; they should be handled by our with
blocks, though.)
# print(repr(root_slicer))
Removing HTML Markup¶
Let us come to our ongoing example, remove_html_markup()
. This is how its instrumented code looks like:
with Slicer(remove_html_markup) as rhm_slicer:
s = remove_html_markup("<foo>bar</foo>")
The graph is as discussed in the introduction to this chapter:
rhm_slicer
# print(repr(rhm_slicer.dependencies()))
rhm_slicer.code()
We can also compute slices over the dependencies:
_, start_remove_html_markup = inspect.getsourcelines(remove_html_markup)
start_remove_html_markup
slicing_criterion = ('tag', (remove_html_markup,
start_remove_html_markup + 9))
tag_deps = rhm_slicer.dependencies().backward_slice(slicing_criterion)
tag_deps
# repr(tag_deps)
Calls and Augmented Assign¶
Our last example covers augmented assigns and data flow across function calls. We introduce two simple functions add_to()
and mul_with()
:
def add_to(n, m):
n += m
return n
def mul_with(x, y):
x *= y
return x
And we put these two together in a single call:
def test_math() -> None:
return mul_with(1, add_to(2, 3))
with Slicer(add_to, mul_with, test_math) as math_slicer:
test_math()
The resulting dependence graph nicely captures the data flow between these calls, notably arguments and parameters:
math_slicer
These are also reflected in the code view:
math_slicer.code()
Dynamic Instrumentation¶
When initializing Slicer()
, one has to provide the set of functions to be instrumented. This is because the instrumentation has to take place before the code in the with
block is executed. Can we determine this list on the fly – while Slicer()
is executed?
The answer is: Yes, but the solution is a bit hackish – even more so than what we have seen above. In essence, we proceed in two steps:
- When
DynamicSlicer.__init__()
is called:- Use the
inspect
module to determine the source code of the call - Analyze the enclosed
with
block for function calls - Instrument these functions
- Use the
- Whenever a function is about to be called (
DataTracker.call()
)- Create an instrumented version of that function
- Have the
call()
method return the instrumented function instead
Implementing Dynamic Instrumentation
We start with the aim of determining the with
block to which our slicer is applied. The our_with_block()
method inspects the source code of the Slicer()
caller, and returns the one with
block (as an AST) whose source line is currently active.
class WithVisitor(NodeVisitor):
def __init__(self) -> None:
self.withs: List[ast.With] = []
def visit_With(self, node: ast.With) -> AST:
self.withs.append(node)
return self.generic_visit(node)
class Slicer(Slicer):
def our_with_block(self) -> ast.With:
"""Return the currently active `with` block."""
frame = self.caller_frame()
source_lines, starting_lineno = inspect.getsourcelines(frame)
starting_lineno = max(starting_lineno, 1)
if len(source_lines) == 1:
# We only get one `with` line, rather than the full block
# This happens in Jupyter notebooks with iPython 8.1.0 and later.
# Here's a hacky workaround to get the cell contents:
# https://stackoverflow.com/questions/51566497/getting-the-source-of-an-object-defined-in-a-jupyter-notebook
source_lines = inspect.linecache.getlines(inspect.getfile(frame))
starting_lineno = 1
source_ast = ast.parse(''.join(source_lines))
wv = WithVisitor()
wv.visit(source_ast)
for with_ast in wv.withs:
if starting_lineno + (with_ast.lineno - 1) == frame.f_lineno:
return with_ast
raise ValueError("Cannot find 'with' block")
Within the with
AST, we can identify all calls:
class CallCollector(NodeVisitor):
def __init__(self) -> None:
self.calls: Set[str] = set()
def visit_Call(self, node: ast.Call) -> AST:
caller_id = ast.unparse(node.func).strip()
self.calls.add(caller_id)
return self.generic_visit(node)
class Slicer(Slicer):
def calls_in_our_with_block(self) -> Set[str]:
"""Return a set of function names called in the `with` block."""
block_ast = self.our_with_block()
cc = CallCollector()
for stmt in block_ast.body:
cc.visit(stmt)
return cc.calls
The method funcs_in_our_with_block()
finally returns a list of all functions used in the with
block. This is the list of functions we will instrument upfront.
class Slicer(Slicer):
def funcs_in_our_with_block(self) -> List[Callable]:
funcs = []
for id in self.calls_in_our_with_block():
func = self.search_func(id)
if func:
funcs.append(func)
return funcs
The method default_items_to_instrument()
is already provided for dynamic instrumentation – it is invoked whenever no list of functions to instrument is provided. So we have it return the list of functions as computed above.
However, default_items_to_instrument()
does one more thing: Using the (also provided) instrument_call()
hook in our dependency tracker, we ensure that all (later) calls are redirected to instrumented functions! This way, if further functions are called, these will be instrumented on the fly.
class Slicer(Slicer):
def default_items_to_instrument(self) -> List[Callable]:
# In _data.call(), return instrumented function
self.dependency_tracker.instrument_call = self.instrument
# Start instrumenting the functions in our `with` block
return self.funcs_in_our_with_block()
Both these hacks are effective, as shown in the following example. We use the Slicer()
constructor without arguments; it automatically identifies fun_2()
as a function in the with
block. As the instrumented fun2()
is invoked, its _data.call()
method instruments the call to fun_1()
(and ensures the instrumented version is called).
def fun_1(x: int) -> int:
return x
def fun_2(x: int) -> int:
return fun_1(x)
with Slicer(log=True) as slicer:
fun_2(10)
slicer
More Applications¶
The main use of dynamic slices is for debugging tools, where they show the origins of individual values. However, beyond facilitating debugging, tracking information flows has a number of additional applications, some of which we briefly sketch here.
Verifying Information Flows¶
Using dynamic slices, we can check all the locations where (potentially sensitive) information is used. As an example, consider the following function password_checker()
, which requests a password from the user and returns True
if it is the correct one:
import hashlib
from bookutils import input, next_inputs
SECRET_HASH_DIGEST = '59f2da35bcc39525b87932b4cc1f3d68'
def password_checker() -> bool:
"""Request a password. Return True if correct."""
secret_password = input("Enter secret password: ")
password_digest = hashlib.md5(secret_password.encode('utf-8')).hexdigest()
if password_digest == SECRET_HASH_DIGEST:
return True
else:
return False
(Note that this is a very naive implementation: A true password checker would use the Python getpass
module to read in a password without echoing it in the clear, and possibly also use a more sophisticated hash function than md5
.)
From a security perspective, the interesting question we can ask using slicing is: Is the entered password stored in the clear somewhere? For this, we can simply run our slicer to see where the inputs are going:
with Slicer() as slicer:
valid_pwd = password_checker()
slicer
We see that the password only flows into password_digest
, where it is already encrypted. If the password were flowing into some other function or variable, we would see this in our slice.
(Note that an attacker may still be able to find out which password was entered, for instance, by checking memory contents.)
Assessing Test Quality¶
Another interesting usage of dynamic slices is to assess test quality. With our square_root()
function, we have seen that the included assertions well test the arguments and the result for correctness:
However, a lazy programmer could also omit these tests – or worse yet, include tests that always pass:
def square_root_unchecked(x):
assert True # <-- new "precondition"
approx = None
guess = x / 2
while approx != guess:
approx = guess
guess = (approx + x / approx) / 2
assert True # <-- new "postcondition"
return approx
How can one check that the tests supplied actually are effective? This is a problem of "Who watches the watchmen" – we need to find a way to ensure that the tests do their job.
The "classical" way of testing tests is so-called mutation testing – that is, introducing artificial errors into the code to see whether the tests catch them. Mutation testing is effective: The above "weak" tests would not catch any change to the square_root()
computation code, and hence quickly be determined as ineffective. However, mutation testing is also costly, as tests have to be ran again and again for every small code mutation.
Slices offer a cost-effective alternative to determine the quality of tests. The idea is that if there are statements in the code whose result does not flow into an assertion, then any errors in these statements will go unnoticed. In consequence, the larger the backward slice of an assertion, the higher its ability to catch errors.
We can easily validate this assumption using the two examples, above. Here is the backward slice for the "full" postcondition in square_root()
. We see that the entire computation code flows into the final postcondition:
postcondition_lineno = start_square_root + 9
postcondition_lineno
with Slicer() as slicer:
y = square_root(4)
slicer.dependencies().backward_slice((square_root, postcondition_lineno))
In contrast, the "lazy" assertion in square_root_unchecked()
has an empty backward slice, showing that it depends on no other value at all:
with Slicer() as slicer:
y = square_root_unchecked(4)
slicer.dependencies().backward_slice((square_root, postcondition_lineno))
In [Schuler et al, 2011], Schuler et al. have tried out this technique and found their "checked coverage" to be a sure indicator for the quality of the checks in tests. Using our dynamic slices, you may wish to try this out on Python code.
Use in Statistical Debugging¶
Collecting dynamic slices over several runs allows for correlating dependencies with other execution features, notably failures: "The program fails whenever the value of weekday
comes from calendar()
." We will revisit this idea in the chapter on statistical debugging.
Things that do not Work¶
Our slicer (and especially the underlying dependency tracker) is still a proof of concept. A number of Python features are not or only partially supported, and/or hardly tested:
- Exceptions can lead to missing or erroneous dependencies. The code assumes that for every
call()
, there is a matchingret()
; when exceptions break this, dependencies across function calls and arguments may be missing or be assigned incorrectly. - Multiple definitions on a single line as in
x = y; x = 1
can lead to missing or erroneous dependencies. Our implementation assumes that there is one statement per line. - If-Expressions (
y = 1 if x else 0
) do not create control dependencies, as there are no statements to control. Neither doif
clauses in comprehensions. - Asynchronous functions (
async
,await
) are not tested.
In these cases, the instrumentation and the underlying dependency tracker may fail to identify control and/or data flows. The semantics of the code, however, should always stay unchanged.
Lessons Learned¶
- To track the origin of some incorrect value, follow back its dependencies:
- Data dependencies indicate where the value came from.
- Control dependencies show why a statement was executed.
- A slice is a subset of the code that could have influenced a specific value. It can be computed by transitively following all dependencies.
- Instrument code to automatically determine and visualize dependencies.
Next Steps¶
In the next chapter, we will explore how to make use of multiple passing and failing executions.
Background¶
Slicing as computing a subset of a program by means of data and control dependencies was invented by Mark Weiser [Weiser et al, 1981]. In his seminal work "Programmers use Slices when Debugging", [Weiser et al, 1982], Weiser demonstrated how such dependencies are crucial for systematic debugging:
When debugging unfamiliar programs programmers use program pieces called slices which are sets of statements related by their flow of data. The statements in a slice are not necessarily textually contiguous, but may be scattered through a program.
Weiser's slices (and dependencies) were determined statically from program code. Both Korel and Laski [Korel et al, 1988] as well as Agrawal and Horgan [Agrawal et al, 1990] introduced dynamic program slicing, building on dynamic dependencies, which would be more specific to a given (failing) run. (The Slicer
we implement in this chapter is a dynamic slicer.) Tip [Tip et al, 1995] gives a survey on program slicing techniques. Chen et al. [Z. Chen et al, 2014] describe and evaluate the first dynamic slicer for Python programs (which is independent of our implementation).
One exemplary application of program slices is the Whyline by Ko and Myers [Ko et al, 2004]. The Whyline is a debugging interface for asking questions about program behavior. It allows querying interactively where a particular variable came from (a data dependency) and why or why not specific things took place (control dependencies).
In [Ezekiel Soremekun et al, 2021], Soremekun et al. evaluated the performance of slicing as a fault localization mechanism and found that following dependencies was one of the most successful strategies to determine fault locations. Notably, if programmers first examine at most the top five most suspicious locations from statistical debugging, and then switch to dynamic slices, on average, they will need to examine only 15% (12 lines) of the code.
Exercises¶
Exercise 1: Control Slices¶
Augment the Slicer
class with two keyword arguments, include
and exclude
, each taking a list of functions to instrument or not to instrument, respectively. These can be helpful when using "automatic" instrumentation.
Exercise 2: Incremental Exploration¶
This is more of a programming project than a simple exercise. Rather than showing all dependencies as a whole, as we do, build a system that allows the user to explore dependencies interactively.
Exercise 3: Forward Slicing¶
Extend Dependencies
with a variant of backward_slice()
named forward_slice()
that, instead of computing the dependencies that go into a location, computes the dependencies that go out of a location.
Exercise 4: Code with Forward Dependencies¶
Create a variant of Dependencies.code()
that, for each statement s
, instead of showing a "passive" view (which variables and locations influenced s
?), shows an "active" view (which variables and locations were influenced by s
?). For middle()
, for instance, the first line should show which lines are influenced by x
, y
, and z
, respectively. Use ->
for control flows and =>
for data flows.
Exercise 5: Flow Assertions¶
In line with Verifying Flows at Runtime, above, implement a function assert_flow(target, source)
that checks at runtime that the data flowing into target
only comes from the variables named in source
.
def assert_flow(target: Any, source: List[Any]) -> bool:
"""
Raise an `AssertionError` if the dependencies of `target`
are not equal to `source`.
"""
...
return True
assert_flow()
would be used in conjunction with Slicer()
as follows:
def demo4() -> int:
x = 25
y = 26
assert_flow(y, [x]) # ensures that `y` depends on `x` only
return y
with Slicer() as slicer:
demo4()
To check dependencies, have assert_flow()
check the contents of the _data
dependency collector as set up by the slicer.
Exercise 6: Checked Coverage¶
Implement checked coverage, as sketched in Assessing Test Quality above. For every assert
statement encountered during a run, produce the number of statements it depends upon.
The content of this project is licensed under the Creative Commons Attribution-NonCommercial-ShareAlike 4.0 International License. The source code that is part of the content, as well as the source code used to format and display that content is licensed under the MIT License. Last change: 2024-11-09 16:52:51+01:00 • Cite • Imprint