ASE BuildSystem ManualMatthewG.KnepleyJuly, 2005Release tag ???IntroductionThe BuildSystem from ASE is intended to be a Python replacement for the GNU autotools. It actually encompasses
somewhat more, as it supports integrated version control and automatic code generation. However, the most useful
comparisons will come from autoconf, make, and libtool. The
system is not designed to be monolithic. Thus each component may be used independently, meaning logging, configuration,
and build are all separate modules which do not require each other. This allows a user to incremenetally adopt the most
useful portions of the package.ConfigureConfigure Design SketchThe system is based upon an autonomous unit, objects of class config.base.Configure, which
are responsible for discovering configuration information for a particular package or purpose. The only interface which
must be supported is the configure method, as shown below. Support for lower-level operations
such as compiling and linking will be discussed in section ???.ConfigureconfigureselfThis collection of configure objects is managed by a config.base.Framework object. As we
will see in section ???, the framework manages all dependecies between modules and output of configure information. The
framework is itself a subclass of config.base.Configure for which the
configure method manages the entire configuration process. In order to associate a module with
the given framework, it also provides the require method, discussed in section ???. Thus, the
minimal framework interface is given by:Frameworkconfig.base.ConfigurerequireselfmoduleNamedepChildkeywordArgs{}configureselfThis design allows user modules to be seamlessly integrated into the framework without changing the paradigm, or
even any of the original code. Modules can be specified on the command line, or left in special directories. Although it
is common to derive from config.base.Configure, the only necessity is that the user provide a
configure method for the framework to execute.The framework does provide the traditional output mechanisms from autoconf, namely
#define statements and file substitutions, to which we add make variables and
rules. However, the preferred interaction mechanism is to use member variables directly from the configure objects. This
is illustrated in section ???Running configureThe first step in running configure is to show the help:
bash$framework.py -help
Python Configure Help
Comma separated lists should be given between [] (use \[ \] in tcsh/csh)
For example: --with-mpi-lib=\[/usr/local/lib/libmpich.a,/usr/local/lib/libpmpich.a\]
----------------------------------------------------------------------------------------
Script:
--help : Print this help message current: 1
--h : Print this help message current: 0
Framework:
--configModules : A list of Python modules with a Configure class current: []
--ignoreCompileOutput : Ignore compiler output current: 1
--ignoreLinkOutput : Ignore linker output current: 1
--ignoreWarnings : Ignore compiler and linker warnings current: 0
--doCleanup : Delete any configure generated files (turn off for debugging) current: 1
--with-alternatives : Provide a choice among alternative package installations current: 0
--with-executables-search-path : A list of directories used to search for executables current: []
--with-packages-search-path : A list of directories used to search for packages current: []
--with-batch : Machine uses a batch system to submit jobs current: 0
The options shown will depend upon the modules loaded with . For instance, we will
normally load the compiler module, which reveals the host of optios controlling preprocessors, compilers, and linkers.
bash$framework.py -configModules=[config.compilers] -help
Python Configure Help
Comma separated lists should be given between [] (use \[ \] in tcsh/csh)
For example: --with-mpi-lib=\[/usr/local/lib/libmpich.a,/usr/local/lib/libpmpich.a\]
----------------------------------------------------------------------------------------
Script:
--help : Print this help message current: 1
--h : Print this help message current: 0
Framework:
--configModules : A list of Python modules with a Configure class current: []
--ignoreCompileOutput : Ignore compiler output current: 1
--ignoreLinkOutput : Ignore linker output current: 1
--ignoreWarnings : Ignore compiler and linker warnings current: 0
--doCleanup : Delete any configure generated files (turn off for debugging) current: 1
--with-alternatives : Provide a choice among alternative package installations current: 0
--with-executables-search-path : A list of directories used to search for executables current: []
--with-packages-search-path : A list of directories used to search for packages current: []
--with-batch : Machine uses a batch system to submit jobs current: 0
Compilers:
--with-cpp=<prog> : Specify the C preprocessor
--with-cc=<prog> : Specify the C compiler
--with-cxx=<prog> : Specify the C++ compiler
--with-fc=<prog> : Specify the Fortran compiler
--with-64-bit-pointers=<bool> : Use 64 bit compilers and libraries current: 0
--CPP=<prog> : Specify the C preprocessor
--CPPFLAGS=<string> : Specify the C preprocessor options current:
--CXXPP=<prog> : Specify the C++ preprocessor
--CC=<prog> : Specify the C compiler
--CFLAGS=<string> : Specify the C compiler options current:
--CXX=<prog> : Specify the C++ compiler
--CXXFLAGS=<string> : Specify the C++ compiler options current:
--CXX_CXXFLAGS=<string> : Specify the C++ compiler-only options current:
--FC=<prog> : Specify the Fortran compiler
--FFLAGS=<string> : Specify the Fortran compiler options current:
--LD=<prog> : Specify the default linker
--CC_LD=<prog> : Specify the linker for C only
--CXX_LD=<prog> : Specify the linker for C++ only
--FC_LD=<prog> : Specify the linker for Fortran only
--LDFLAGS=<string> : Specify the linker options current:
--with-ar : Specify the archiver
-AR : Specify the archiver flags
-AR_FLAGS : Specify the archiver flags
--with-ranlib : Specify ranlib
--with-shared-libraries : Enable shared libraries current: 1
--with-shared-ld=<prog> : Specify the shared linker
--with-f90-header=<file> : Specify the C header for the F90 interface, e.g. f90_intel.h
--with-f90-source=<file> : Specify the C source for the F90 interface, e.g. f90_intel.c
The syntax for list and dictionary option values is identical to Python syntax. However, in some shells (like
csh), brackets must be escaped, and braces will usually have to be enclosed in quotes.The modules indicated with are located using PYTHONPATH. Since
specifying environment variables can be inconvenient and error prone, it is common to provide a driver which alters
sys.path, as is done for PETSc. In fact, the PETSc driver
Verifies PETSC_ARCHChecks for invalid Cygwin versionsChecks for RedHat 9, which has a threads bugAugments PYTHONPATHAdds the default PETSc configure modulePersists the configuration in RDict.dbHandles exceptionsAdding a moduleAs we discussed in the introduction, all that is strictly necessary for a configure module, is to provide a class
named Configure with a method configure taking no arguments. However,
there are a variety of common operations, which will be illustrated in the sections below.Using other modulesWe will often want to use the methods or results of other configure modules in order to perform checks in our
own. The framework provides a mechanism for retrieving the object for any given configure module. As an example,
consider checking for the ddot function in the BLAS library. The relevant Python code would
be
import config.base
class Configure(config.base.Configure):
def __init__(self, framework):
config.base.Configure.__init__(self, framework)
self.compilers = self.framework.require('config.compilers', self)
self.libraries = self.framework.require('config.libraries', self)
return
def configure(self):
return self.libraries.check('libblas.a', 'ddot', otherLibs = self.compilers.flibs,
fortranMangle = 1)
The require call will return the configure object from the given module, creating it if
necessary. If the second argument is given, the framework will ensure that the returned configure object runs
before the passed configure object. Notice that we can use the returned object either to call
methods, like check from config.libraries, or use member variables, such
as the list of Fortran compatibility libraries flibs from
config.compilers.
The underlying implementation in the framework uses a directed acyclic graph to indicate dependencies among
modules. The vertices of this graph, configure objects, are topologically sorted and then executed. Moreover, child
objects can be added to the framework without respecting the dependency structure, but this is discouraged.Adding a testA user could of course perform all tests in the object's configure method, but the base
class provides useful logging support for this purpose. Consider again the BLAS example, which will now become,
def checkDot(self):
'''Verify that the ddot() function is contained in the BLAS library'''
return self.libraries.check('libblas.a', 'ddot', otherLibs = self.compilers.flibs,
fortranMangle = 1)
def configure(self):
self.executeTest(self.checkDot)
return
Passing our test module to the framework,
docs$PYTHONPATH=`pwd` ../config/framework.py --configModules=[examples.blasTest]
we produce the following log output in configure.log. Notice that it not only records the method and module, but the method doc string,
all shell calls, and any output actions as well.
================================================================================
TEST checkDot from examples.blasTest(/PETSc3/sidl/BuildSystem/docs/examples/blasTest.py:10)
TESTING: checkDot from examples.blasTest(/PETSc3/sidl/BuildSystem/docs/examples/blasTest.py:10)
Verify that the ddot() function is contained in the BLAS library
Checking for functions ['ddot'] in library ['libblas.a'] ['-lfrtbegin', '-lg2c', '-lm',
'-L/usr/lib/gcc-lib/i486-linux/3.3.5', '-L/usr/lib/gcc-lib/i486-linux/3.3.5/../../..',
'-lm', '-lgcc_s']
sh: gcc -c -o conftest.o -fPIC conftest.c
Executing: gcc -c -o conftest.o -fPIC conftest.c
sh:
sh: gcc -o conftest -fPIC conftest.o -lblas -lfrtbegin -lg2c -lm
-L/usr/lib/gcc-lib/i486-linux/3.3.5 -L/usr/lib/gcc-lib/i486-linux/3.3.5/../../.. -lm -lgcc_s
Executing: gcc -o conftest -fPIC conftest.o -lblas -lfrtbegin -lg2c -lm
-L/usr/lib/gcc-lib/i486-linux/3.3.5 -L/usr/lib/gcc-lib/i486-linux/3.3.5/../../.. -lm -lgcc_s
sh:
Defined HAVE_LIBBLAS to 1 in config.libraries
Checking for headersOften, we would like to test for the presence of certain headers. This is done is a completely analogous way to
the library case, using instead the config.headers module. Below, we test for the presence of the
curses header.
import config.base
class Configure(config.base.Configure):
def __init__(self, framework):
config.base.Configure.__init__(self, framework)
self.headers = self.framework.require('config.headers, self)
return
def checkCurses(self):
'Verify that we have the curses header'
return self.headers.check('curses.h')
def configure(self):
self.executeTest(self.checkCurses)
return
Running this test
docs$PYTHONPATH=`pwd` ../config/framework.py --configModules=[examples.cursesTest]
produces the following log output.
================================================================================
TEST checkCurses from examples.cursesTest(/PETSc3/sidl/BuildSystem/docs/examples/cursesTest.py:9)
TESTING: checkCurses from examples.cursesTest(/PETSc3/sidl/BuildSystem/docs/examples/cursesTest.py:9)
Verify that we have the curses header
Checking for header: curses.h
sh: gcc -E conftest.c
Executing: gcc -E conftest.c
sh: # 1 "conftest.c"
# 1 "<built-in>"
# 1 "<command line>"
# 1 "conftest.c"
# 1 "confdefs.h" 1
# 2 "conftest.c" 2
# 1 "conffix.h" 1
# 3 "conftest.c" 2
# 1 "/usr/include/curses.h" 1 3 4
# 58 "/usr/include/curses.h" 3 4
# 1 "/usr/include/ncurses_dll.h" 1 3 4
# 59 "/usr/include/curses.h" 2 3 4
# 99 "/usr/include/curses.h" 3 4
typedef unsigned long chtype;
# 1 "/usr/include/stdio.h" 1 3 4
# 28 "/usr/include/stdio.h" 3 4
# 1 "/usr/include/features.h" 1 3 4
# 295 "/usr/include/features.h" 3 4
# 1 "/usr/include/sys/cdefs.h" 1 3 4
# 296 "/usr/include/features.h" 2 3 4
# 318 "/usr/include/features.h" 3 4
#...
... W* win,int* y, int* x, _Bool to_screen);
extern _Bool mouse_trafo (int*, int*, _Bool);
extern int mcprint (char *, int);
extern int has_key (int);
extern void _tracef (const char *, ...) ;
extern void _tracedump (const char *, WINDOW *);
extern char * _traceattr (attr_t);
extern char * _traceattr2 (int, chtype);
extern char * _nc_tracebits (void);
extern char * _tracechar (int);
extern char * _tracechtype (chtype);
extern char * _tracechtype2 (int, chtype);
# 1203 "/usr/include/curses.h" 3 4
extern char * _tracemouse (const MEVENT *);
extern void trace (const unsigned int);
# 4 "conftest.c" 2
Defined HAVE_CURSES_H to 1 in config.headers
Alternatively, we could have specified that this header be included in the list of header files checked by default.
import config.base
class Configure(config.base.Configure):
def __init__(self, framework):
config.base.Configure.__init__(self, framework)
self.headers = self.framework.require('config.headers, self)
self.headers.headers.append('curses.h')
return
def checkCurses(self):
'Verify that we have the curses header'
return self.headers.haveHeader('curses.h')
def configure(self):
self.executeTest(self.checkCurses)
return
In addition, the base class does include lower level support for preprocessing files. The
preprocess method takes a code string as input and return a tuple of the
(stdout,stderr,error code) for the run. The outputPreprocess method returns
only the standard output, and checkPreprocess returns true if no error occurs.Checking for librariesWe have already demonstrated a test for the existence of a function in a library. However the
check method is much more general. It allows the specification of multiple libraries and
multiple functions, as well as auxiliary libraries. For instance, to check for the MPI_Init and
MPI_Comm_create functions in MPICH when the Fortran bindings are active, we would use:
self.libraries.check(['libmpich.so', 'libpmpich.so'], ['MPI_Init', 'MPI_Comm_create'],
otherLibs = self.compilers.flibs)
As in the BLAS example, we can also turn on Fortran name mangling. The caller may also supply a function prototype and
calling sequence, which are necessary if the current language is C++.
It is also necessary at some times to determine whether a given library is a shared object. This can be
accomplished using the checkShared method, as we demonstrate with the MPICH library in a call
taken from the MPI configure module in PETSc.
self.libraries.checkShared('#include <mpi.h>\n', 'MPI_Init', 'MPI_Initialized',
'MPI_Finalize', checkLink = self.checkMPILink,
libraries = self.lib)
The theory for the check is that a shared object will have only one copy of any global variable. Thus functions such as
MPI_Initialized will render consistent results across other libraries. The test begins by
creating two dynamic libraries, both of which link the given library. Then an executable is constructed which loads the
libraries in turn. The first library calls the initizlization functions, here MPI_Init, and the
second library calls the initialization check function, here MPI_Initialized. The check
function will return true if the given library is a shared object. This organization is shown in figure ???The lower level interface to compiling and linking in the base class mirrors that for preprocessing. The
outputCompile and checkCompile methods function in the same way. The
code is now broken up into four distinct sections. There are includes, the body of main, and a
possible replacement for the beginning and end of the main declaration. The linking methods,
outputLink and checkLink, are exactly analogous.There are also some convenience methods provided to handle compiler and linker flags. The
checkCompilerFlag and checkLinkerFlag try to determine whether a given
flag is accepted by the processor, while addCompilerFlag and
addLinkerFlag will do that check and add any valid flag to the list of default flags.Checking for executablesThe getExecutable method is used to locate executable files. For instance, this code
would allow us to locate the valgrind binary.
self.getExecutable('valgrind')
If the program is found, a member variable of the same name will be set in the object to the program name, and a make
macro defined to it as well. We can opt for these to contain the full path by using the
argument. In addition, we can change the name of the member variable and macro using the
argument.
We also have control over the search path used. If we give no arguments, the default path from the environment is
used. This can be overridden with a new path using the argument, either as a Python list or a
colon separated string. Furthermore, the default path can be added to this custom path using the
argument. For instance, this call
self.getExecutable('valgrind', path=['/opt/valgrind-1.0'], getFullPath=1,
useDefaultPath=1, resultName='grinder')
will check for valgrind first in /opt/valgrind-1.0 and then along the default
path. If found in the first location, it will set self.grinder to
/opt/valgrind-1.0/valgrind as well as define GRINDER to the same value in makefiles.
As in the cases of preprocessing, compiling, and linking, the lower level operations are also exposed. The
checkRun method takes in a code string and returns true if the executable runs without
error. The outputRun method returns the output and status code. Both methods us the safe
execution routine config.base.Configure.executeShellCommand which accepts a timeout. Moreover,
there commands can run in the special batch mode described in section ???.Output resultsThe BuildSystem configure includes the traditional output methods employed by autoconf to
enable communication with make. Individual configure modules use the
addDefine method to add C #define statements to a configuration header
and the addSubstitution to setup substitution rules for specified files. For instance, to
activate the parmetis package, we might provide
self.addDefine('HAVE_PARMETIS', 1)
and then for the make process
self.addSubstitution('PARMETIS_INCLUDE', ' '.join([self.libraries.getIncludeArgument(i)
for i in self.include]))
self.addSubstitution('PARMETIS_LIB, ' '.join([self.libraries.getLibArgument(l)
for l in self.lib]))
The actual output of this data is controlled by the framework. The user specifies the header file using the
header field of the framework, and then the file is created automatically during the configure
process, but can be output at any time using the outputHeader method. Furthermore, the
addSubstitutionFile method can be used to tag a file for substitution, and also specify a
different file for the result of the substitution.In the autoconf approach, separating the defines and substitutions for different packages
becomes troublesome, and in some cases impossible to maintain. To help with this, we have introduced
prefixes for the defines and substitutions. The are strings, unique to each module, which are
prepended with an underscore to each identifier defined or substituted. These are set on a per object basis using the
headerPrefix and substPrefix members. For instance, in our
parmetis example, if we instead used the code
self.headerPrefix = 'MATT'
self.addDefine('HAVE_PARMETIS', 1)
in our configuration header we would see
#ifndef MATT_HAVE_PARMETIS
#define MATT_HAVE_PARMETIS 1
#endif
Note that the value of the prefix is used at output time, not at the time that the define or substitution is set.
Another extension of the old-style output mechanisms adds more C structure to the interface. The
addTypedef method allows a typedef from one typename to another, which in
autoconf is handled by a define. Likewise addPrototype can add a missing
function prototype to a header. Since these are C specific structures, they are output into a separate configuration
header file, which is controlled by the cHeader member variable.Extending in a different direction, we allow makefile structures to be specified directly rather than through
substitutions. Using addMakeMacro, we can add variable definitions to the configuration
makefile, whereas addMakeRule allows the user to specify a make target, complete with
dependencies and action. As an example, we will replace our parmetis example from above with the following code
self.addMakeMacro('PARMETIS_INCLUDE', ' '.join([self.libraries.getIncludeArgument(i)
for i in self.include]))
self.addMakeMacro('PARMETIS_LIB, ' '.join([self.libraries.getLibArgument(l)
for l in self.lib]))
self.addMakeRule('.c.o', '', ['${CC} -c -o $@ -I${PARMETIS_INCLUDE} $<'])
self.addMakeRule('myApp', '${.c=.o:SOURCE}', ['${CC} -o $@ $< ${PARMETIS_LIB}'])
which will produce
PARMETIS_INCLUDE = -I/home/knepley/petsc-dev/externalpackages/parmetis/build/Darwin-x86_64/include
PARMETIS_LIB = -L/home/knepley/petsc-dev/externalpackages/parmetis/build/Darwin-x86_64/lib -lparmetis -lmetis
in the file specified by the makeMacroHeader member variable, and
myApp: ${.c=.o:SOURCE}
${CC} -i $@ $< ${PARMETIS_LIB}
in the file specified by the makeRuleHeader member variable.The above output methods are all specified on a per configure object basis, however this may become confusing in a
large project. All the prefixes and output filenames would have to be coordinated. A common strategy is to use the
framework for coordination, putting all the output into the framework object itself. For instance, we might have
self.framework.addDefine('HAVE_PARMETIS', 1)
which would allow the define to appear in the headre specified by the framework with the framework prefix.
Configuring batch systemsIt is not uncommon for large clusters or supercomputing centers to have a batch execution policy, making it
difficult for configure to execute the few tests that depend on executing code, rather than compiling and linking it. To
handle this case, we provide the argument. The code to be run is collected in a single
executable which the user must submit to the system. This executable produces a reconfigure script
which may then be run to fully configure the system.When configure is run with the option, the following message will appear.
petsc-dev$./config/configure.py --with-batch
produces the following log output.
=================================================================================
Since your compute nodes require use of a batch system or mpiexec you must:
1) Submit ./conftest to your batch system (this will generate the file reconfigure)
2) Run "python reconfigure" (to complete the configure process).
=================================================================================
The user must then execute the conftest binary, and then run the python
reconfigure command.
If a user defined test relies upon running code, he may make it suitable for a batch system. The
checkRun method takes the argument which names a configure option
whose value may substitute for the outcome of the test, allowing a user to preempt the run. For instance, the
config.types.checkEndian method contains the code
if self.checkRun('', body, defaultArg = 'isLittleEndian'):
which means the option can be given to replace the output of the run. However, this does
the require the user to supply the missing option.To automate this process, the test should first check for batch mode. Using the
addBatchInclude and addBatchBody methods, code can be included in the
batch executable. We return to the endian test to illustrate this usage.
if not self.framework.argDB['with-batch']:
body = '''
/* Are we little or big endian? From Harbison & Steele. */
union
{
long l;
char c[sizeof(long)];
} u;
u.l = 1;
exit(u.c[sizeof(long) - 1] == 1);
'''
if self.checkRun('', body, defaultArg = 'isLittleEndian'):
endian = 'little'
else:
endian = 'big'
else:
self.framework.addBatchBody(
['{',
' union {long l; char c[sizeof(long)];} u;',
' u.l = 1;',
' fprintf(output, " \'--with-endian=%s\',\\n",\
(u.c[sizeof(long) - 1] == 1) ? "little" : "big");',
'}'])
# Dummy value
endian = 'little'
The batch body code should output configure options to the output file descriptor. These are
collected for the new configure run in the reconfigure script.
BuildThe build operation now encompasses the configure, compile, link, install, and update operations.Running makeAll options for both configuration and build are given to make.py. Thus, the simplest build
is merely
petsc-dev$./make.py
The help is also given by , but this time it will also include build switches.
petsc-dev$./make.py -help
Script Help
-----------
Script:
--help : Print this help message current: 1
--h : Print this help message current: 0
Make:
-forceConfigure : Force a reconfiguration current: 0
-ignoreCompileOutput : Ignore compiler output current: 1
-defaultRoot : Directory root for all packages current: ../..
-prefix : Root for installation of libraries and binaries
SIDLMake:
-bootstrap : Generate the bootstrap client current: 0
-outputSIDLFiles : Write generated files to disk current: 1
-excludeLanguages=<languages> : Do not load configurations from RDict for the given languages current: []
-excludeBasenames=<names> : Do not load configurations from RDict for these SIDL base names current: []
MakersThe build operation now encompasses configure, compile, and link operations, which are coordinated by objects of
class maker.Maker. This object manages:
configuration,build,install, andproject dependencies
All options, no matter which component they are intended for, are given uniformly to make.py.
SIDLMakerThis is a subclass which handles source generation from SIDL.BuildersThe build operation now encompasses the configure, compile, and link operations.LanguageProcessorsThe build operation now encompasses the configure, compile, and link operations.Interaction with ConfigureThe pickled configure is loaded by Maker, and then the config.compile objects are jacked into the Builder.