Although causality is still a highly debated issue among
we believe it provides a natural and intuitive representation about
relations among variables. We propose the use of causal models for
modeling how variables influence the overall performance of
The following figure represents a simplified performance model of a
It contains application parameters (at the left), system parameters
the bottom), the overall performance (at the right) and intermediate
variables that can be measured inside the application of the system.
The following figure shows the basic causal model of the performance
a parallel application:
Besides the formalization of causality by Judea Pearl,
important scientific advance was the invention of algorithms that
learn causal models from observational data. The research of
Glymour and Scheines (SGS) has resulted in the TETRAD tool.
work is based on TETRAD's algorithms, which we extended to
capture the full complexity of performance models:
A mixture of continuous and discrete data.
A form-free measure of dependency.
The handling of deterministic relations.
Context-specific variables and independencies.
The first two problems are solved by using the mutual information as
dependency measure, which is based on Shannon's entropy of
distributions. Kernel Density Estimation is a technique for deriving
distribution from experimental data. The development was conducted
The handling of deterministic relations is attained by considering
case that two variables have equivalent
information about a third variable. The only objective
for deciding upon which variables are directly related is then the
complexity of the relations. This work is done by Jan
Lemeire. The last extension is currently being studied by
Borms, he will use context-specific graphs. Our causal
learning module Research
results, explained with examples
Philosophy of Learning, Minimum
Length & Assumptions
One of the hardest criticisms of causality and the learning
are the assumptions that are made.
We want to answer by considering learning in general, in a
philosophical way: what can we learn from observations? It is clear
that no method can guarantee that the 'real' model will be learnt.
An interesting side question is if the real, physical world can be
known. After all, everything we know about the world comes from
observation (and partly from a book). See the opinions from Jorma Rissanen
and Peter Grünwald about
By linking the work of causal structure learning with the ideas
developed by the field of Algorithmic Information Theory (see Paul Vitanyi) around
regularities (= compression!) and typicalness (randomness =
incompressibility), I believe that we an conclude that we have to
search for regularities in data and construct models that capture
regularities. Moreover, causal models suggest that we have to look
models that capture all