/setup.py
http://unladen-swallow.googlecode.com/ · Python · 1966 lines · 1338 code · 220 blank · 408 comment · 359 complexity · 876f60a7a1997b71b1e8eb54e199eaf4 MD5 · raw file
Large files are truncated click here to view the full file
- # Autodetecting setup.py script for building the Python extensions
- #
- __version__ = "$Revision: 75282 $"
- import sys, os, imp, re, optparse
- from glob import glob
- from platform import machine as platform_machine
- from distutils import log
- from distutils import sysconfig
- from distutils import text_file
- from distutils.errors import *
- from distutils.core import Extension, setup
- from distutils.command.build_ext import build_ext
- from distutils.command.install import install
- from distutils.command.install_lib import install_lib
- from distutils.util import get_platform
- # This global variable is used to hold the list of modules to be disabled.
- disabled_module_list = []
- def record_build_dir():
- """Record the relative path to the build diretory for site.py."""
- build_dir = "build/lib.%s-%.3s" % (get_platform(), sys.version)
- build_dir_file = os.path.join(os.path.dirname(sys.executable), "build_dir")
- f = open(build_dir_file, "w")
- try:
- f.write(build_dir)
- finally:
- f.close()
- def add_dir_to_list(dirlist, dir):
- """Add the directory 'dir' to the list 'dirlist' (at the front) if
- 1) 'dir' is not already in 'dirlist'
- 2) 'dir' actually exists, and is a directory."""
- if dir is not None and os.path.isdir(dir) and dir not in dirlist:
- dirlist.insert(0, dir)
- def find_file(filename, std_dirs, paths):
- """Searches for the directory where a given file is located,
- and returns a possibly-empty list of additional directories, or None
- if the file couldn't be found at all.
- 'filename' is the name of a file, such as readline.h or libcrypto.a.
- 'std_dirs' is the list of standard system directories; if the
- file is found in one of them, no additional directives are needed.
- 'paths' is a list of additional locations to check; if the file is
- found in one of them, the resulting list will contain the directory.
- """
- # Check the standard locations
- for dir in std_dirs:
- f = os.path.join(dir, filename)
- if os.path.exists(f): return []
- # Check the additional directories
- for dir in paths:
- f = os.path.join(dir, filename)
- if os.path.exists(f):
- return [dir]
- # Not found anywhere
- return None
- def find_library_file(compiler, libname, std_dirs, paths):
- result = compiler.find_library_file(std_dirs + paths, libname)
- if result is None:
- return None
- # Check whether the found file is in one of the standard directories
- dirname = os.path.dirname(result)
- for p in std_dirs:
- # Ensure path doesn't end with path separator
- p = p.rstrip(os.sep)
- if p == dirname:
- return [ ]
- # Otherwise, it must have been in one of the additional directories,
- # so we have to figure out which one.
- for p in paths:
- # Ensure path doesn't end with path separator
- p = p.rstrip(os.sep)
- if p == dirname:
- return [p]
- else:
- assert False, "Internal error: Path not found in std_dirs or paths"
- def module_enabled(extlist, modname):
- """Returns whether the module 'modname' is present in the list
- of extensions 'extlist'."""
- extlist = [ext for ext in extlist if ext.name == modname]
- return len(extlist)
- def find_module_file(module, dirlist):
- """Find a module in a set of possible folders. If it is not found
- return the unadorned filename"""
- list = find_file(module, [], dirlist)
- if not list:
- return module
- if len(list) > 1:
- log.info("WARNING: multiple copies of %s found"%module)
- return os.path.join(list[0], module)
- def combine_dirs_to_check(extra_dirs, orig_lib_dirs):
- extra_dirs = [d for d in extra_dirs if d and os.path.isdir(d)]
- # First search extra directories that already appear in the
- # original list.
- extra_dirs.sort(key=lambda d:0 if d in orig_lib_dirs else 1)
- return extra_dirs + orig_lib_dirs
- class PyBuildExt(build_ext):
- def __init__(self, dist):
- build_ext.__init__(self, dist)
- self.failed = []
- def build_extensions(self):
- # Detect which modules should be compiled
- missing = self.detect_modules()
- # Remove modules that are present on the disabled list
- extensions = [ext for ext in self.extensions
- if ext.name not in disabled_module_list]
- # move ctypes to the end, it depends on other modules
- ext_map = dict((ext.name, i) for i, ext in enumerate(extensions))
- if "_ctypes" in ext_map:
- ctypes = extensions.pop(ext_map["_ctypes"])
- extensions.append(ctypes)
- self.extensions = extensions
- # Fix up the autodetected modules, prefixing all the source files
- # with Modules/ and adding Python's include directory to the path.
- (srcdir,) = sysconfig.get_config_vars('srcdir')
- if not srcdir:
- # Maybe running on Windows but not using CYGWIN?
- raise ValueError("No source directory; cannot proceed.")
- # Figure out the location of the source code for extension modules
- # (This logic is copied in distutils.test.test_sysconfig,
- # so building in a separate directory does not break test_distutils.)
- moddir = os.path.join(os.getcwd(), srcdir, 'Modules')
- moddir = os.path.normpath(moddir)
- srcdir, tail = os.path.split(moddir)
- srcdir = os.path.normpath(srcdir)
- moddir = os.path.normpath(moddir)
- moddirlist = [moddir]
- incdirlist = ['./Include']
- # Platform-dependent module source and include directories
- platform = self.get_platform()
- if platform in ('darwin', 'mac') and ("--disable-toolbox-glue" not in
- sysconfig.get_config_var("CONFIG_ARGS")):
- # Mac OS X also includes some mac-specific modules
- macmoddir = os.path.join(os.getcwd(), srcdir, 'Mac/Modules')
- moddirlist.append(macmoddir)
- incdirlist.append('./Mac/Include')
- alldirlist = moddirlist + incdirlist
- # Fix up the paths for scripts, too
- self.distribution.scripts = [os.path.join(srcdir, filename)
- for filename in self.distribution.scripts]
- # Python header files
- headers = glob("Include/*.h") + ["pyconfig.h"]
- for ext in self.extensions[:]:
- ext.sources = [ find_module_file(filename, moddirlist)
- for filename in ext.sources ]
- if ext.depends is not None:
- ext.depends = [find_module_file(filename, alldirlist)
- for filename in ext.depends]
- else:
- ext.depends = []
- # re-compile extensions if a header file has been changed
- ext.depends.extend(headers)
- ext.include_dirs.append( '.' ) # to get config.h
- for incdir in incdirlist:
- ext.include_dirs.append( os.path.join(srcdir, incdir) )
- # If a module has already been built statically,
- # don't build it here
- if ext.name in sys.builtin_module_names:
- self.extensions.remove(ext)
- if platform != 'mac':
- # Parse Modules/Setup and Modules/Setup.local to figure out which
- # modules are turned on in the file.
- remove_modules = []
- for filename in ('Modules/Setup', 'Modules/Setup.local'):
- input = text_file.TextFile(filename, join_lines=1)
- while 1:
- line = input.readline()
- if not line: break
- line = line.split()
- remove_modules.append(line[0])
- input.close()
- for ext in self.extensions[:]:
- if ext.name in remove_modules:
- self.extensions.remove(ext)
- # When you run "make CC=altcc" or something similar, you really want
- # those environment variables passed into the setup.py phase. Here's
- # a small set of useful ones.
- compiler = os.environ.get('CC')
- args = {}
- # unfortunately, distutils doesn't let us provide separate C and C++
- # compilers
- if compiler is not None:
- (ccshared,cflags) = sysconfig.get_config_vars('CCSHARED','CFLAGS')
- args['compiler_so'] = compiler + ' ' + ccshared + ' ' + cflags
- self.compiler.set_executables(**args)
- build_ext.build_extensions(self)
- longest = max([len(e.name) for e in self.extensions])
- if self.failed:
- longest = max(longest, max([len(name) for name in self.failed]))
- def print_three_column(lst):
- lst.sort(key=str.lower)
- # guarantee zip() doesn't drop anything
- while len(lst) % 3:
- lst.append("")
- for e, f, g in zip(lst[::3], lst[1::3], lst[2::3]):
- print "%-*s %-*s %-*s" % (longest, e, longest, f,
- longest, g)
- if missing:
- print
- print "Failed to find the necessary bits to build these modules:"
- print_three_column(missing)
- print ("To find the necessary bits, look in setup.py in"
- " detect_modules() for the module's name.")
- print
- if self.failed:
- failed = self.failed[:]
- print
- print "Failed to build these modules:"
- print_three_column(failed)
- print
- def build_extension(self, ext):
- if ext.name == '_ctypes':
- if not self.configure_ctypes(ext):
- return
- try:
- build_ext.build_extension(self, ext)
- except (CCompilerError, DistutilsError), why:
- self.announce('WARNING: building of extension "%s" failed: %s' %
- (ext.name, sys.exc_info()[1]))
- self.failed.append(ext.name)
- return
- # Workaround for Mac OS X: The Carbon-based modules cannot be
- # reliably imported into a command-line Python
- if 'Carbon' in ext.extra_link_args:
- self.announce(
- 'WARNING: skipping import check for Carbon-based "%s"' %
- ext.name)
- return
- if self.get_platform() == 'darwin' and (
- sys.maxint > 2**32 and '-arch' in ext.extra_link_args):
- # Don't bother doing an import check when an extension was
- # build with an explicit '-arch' flag on OSX. That's currently
- # only used to build 32-bit only extensions in a 4-way
- # universal build and loading 32-bit code into a 64-bit
- # process will fail.
- self.announce(
- 'WARNING: skipping import check for "%s"' %
- ext.name)
- return
- # Workaround for Cygwin: Cygwin currently has fork issues when many
- # modules have been imported
- if self.get_platform() == 'cygwin':
- self.announce('WARNING: skipping import check for Cygwin-based "%s"'
- % ext.name)
- return
- ext_filename = os.path.join(
- self.build_lib,
- self.get_ext_filename(self.get_ext_fullname(ext.name)))
- try:
- imp.load_dynamic(ext.name, ext_filename)
- except ImportError, why:
- self.failed.append(ext.name)
- self.announce('*** WARNING: renaming "%s" since importing it'
- ' failed: %s' % (ext.name, why), level=3)
- assert not self.inplace
- basename, tail = os.path.splitext(ext_filename)
- newname = basename + "_failed" + tail
- if os.path.exists(newname):
- os.remove(newname)
- os.rename(ext_filename, newname)
- # XXX -- This relies on a Vile HACK in
- # distutils.command.build_ext.build_extension(). The
- # _built_objects attribute is stored there strictly for
- # use here.
- # If there is a failure, _built_objects may not be there,
- # so catch the AttributeError and move on.
- try:
- for filename in self._built_objects:
- os.remove(filename)
- except AttributeError:
- self.announce('unable to remove files (ignored)')
- except:
- exc_type, why, tb = sys.exc_info()
- self.announce('*** WARNING: importing extension "%s" '
- 'failed with %s: %s' % (ext.name, exc_type, why),
- level=3)
- self.failed.append(ext.name)
- def get_platform(self):
- # Get value of sys.platform
- for platform in ['cygwin', 'beos', 'darwin', 'atheos', 'osf1']:
- if sys.platform.startswith(platform):
- return platform
- return sys.platform
- def detect_modules(self):
- use_system_paths = not os.getenv('IGNORE_SYSTEM_PATHS')
- if use_system_paths:
- # Ensure that /usr/local is always used
- add_dir_to_list(self.compiler.library_dirs, '/usr/local/lib')
- add_dir_to_list(self.compiler.include_dirs, '/usr/local/include')
- # Add paths specified in the environment variables LDFLAGS and
- # CPPFLAGS for header and library files.
- # We must get the values from the Makefile and not the environment
- # directly since an inconsistently reproducible issue comes up where
- # the environment variable is not set even though the value were passed
- # into configure and stored in the Makefile (issue found on OS X 10.3).
- for env_var, arg_name, dir_list in (
- ('LDFLAGS', '-R', self.compiler.runtime_library_dirs),
- ('LDFLAGS', '-L', self.compiler.library_dirs),
- ('CPPFLAGS', '-I', self.compiler.include_dirs)):
- env_val = sysconfig.get_config_var(env_var)
- if env_val:
- # To prevent optparse from raising an exception about any
- # options in env_val that it doesn't know about we strip out
- # all double dashes and any dashes followed by a character
- # that is not for the option we are dealing with.
- #
- # Please note that order of the regex is important! We must
- # strip out double-dashes first so that we don't end up with
- # substituting "--Long" to "-Long" and thus lead to "ong" being
- # used for a library directory.
- env_val = re.sub(r'(^|\s+)-(-|(?!%s))' % arg_name[1],
- ' ', env_val)
- parser = optparse.OptionParser()
- # Make sure that allowing args interspersed with options is
- # allowed
- parser.allow_interspersed_args = True
- parser.error = lambda msg: None
- parser.add_option(arg_name, dest="dirs", action="append")
- options = parser.parse_args(env_val.split())[0]
- if options.dirs:
- for directory in reversed(options.dirs):
- add_dir_to_list(dir_list, directory)
- if os.path.normpath(sys.prefix) != '/usr':
- add_dir_to_list(self.compiler.library_dirs,
- sysconfig.get_config_var("LIBDIR"))
- add_dir_to_list(self.compiler.include_dirs,
- sysconfig.get_config_var("INCLUDEDIR"))
- try:
- have_unicode = unicode
- except NameError:
- have_unicode = 0
- # lib_dirs and inc_dirs are used to search for files;
- # if a file is found in one of those directories, it can
- # be assumed that no additional -I,-L directives are needed.
- lib_dirs = list(self.compiler.library_dirs)
- inc_dirs = list(self.compiler.include_dirs)
- if use_system_paths:
- lib_dirs += [
- '/lib64', '/usr/lib64',
- '/lib', '/usr/lib',
- ]
- inc_dirs += ['/usr/include']
- exts = []
- missing = []
- config_h = sysconfig.get_config_h_filename()
- config_h_vars = sysconfig.parse_config_h(open(config_h))
- platform = self.get_platform()
- (srcdir,) = sysconfig.get_config_vars('srcdir')
- # Check for AtheOS which has libraries in non-standard locations
- if platform == 'atheos':
- if use_system_paths:
- lib_dirs += ['/system/libs', '/atheos/autolnk/lib']
- lib_dirs += os.getenv('LIBRARY_PATH', '').split(os.pathsep)
- inc_dirs += ['/system/include', '/atheos/autolnk/include']
- inc_dirs += os.getenv('C_INCLUDE_PATH', '').split(os.pathsep)
- # OSF/1 and Unixware have some stuff in /usr/ccs/lib (like -ldb)
- if (platform in ['osf1', 'unixware7', 'openunix8'] and
- use_system_paths):
- lib_dirs += ['/usr/ccs/lib']
- if platform == 'darwin' or platform == 'linux2':
- # This should work on any unixy platform ;-)
- # If the user has bothered specifying additional -I and -L flags
- # in OPT and LDFLAGS we might as well use them here.
- # NOTE: using shlex.split would technically be more correct, but
- # also gives a bootstrap problem. Let's hope nobody uses directories
- # with whitespace in the name to store libraries.
- cflags, ldflags = sysconfig.get_config_vars(
- 'CFLAGS', 'LDFLAGS')
- for item in cflags.split():
- if item.startswith('-I'):
- inc_dirs.append(item[2:])
- for item in ldflags.split():
- if item.startswith('-L'):
- lib_dirs.append(item[2:])
- # Check for MacOS X, which doesn't need libm.a at all
- math_libs = ['m']
- if platform in ['darwin', 'beos', 'mac']:
- math_libs = []
- # XXX Omitted modules: gl, pure, dl, SGI-specific modules
- #
- # The following modules are all pretty straightforward, and compile
- # on pretty much any POSIXish platform.
- #
- # Some modules that are normally always on:
- exts.append( Extension('_weakref', ['_weakref.c']) )
- # array objects
- exts.append( Extension('array', ['arraymodule.c']) )
- # complex math library functions
- exts.append( Extension('cmath', ['cmathmodule.c'],
- libraries=math_libs) )
- # math library functions, e.g. sin()
- exts.append( Extension('math', ['mathmodule.c'],
- libraries=math_libs) )
- # fast string operations implemented in C
- exts.append( Extension('strop', ['stropmodule.c']) )
- # time operations and variables
- exts.append( Extension('time', ['timemodule.c'],
- libraries=math_libs) )
- exts.append( Extension('datetime', ['datetimemodule.c', 'timemodule.c'],
- libraries=math_libs) )
- # fast iterator tools implemented in C
- exts.append( Extension("itertools", ["itertoolsmodule.c"]) )
- # code that will be builtins in the future, but conflict with the
- # current builtins
- exts.append( Extension('future_builtins', ['future_builtins.c']) )
- # random number generator implemented in C
- exts.append( Extension("_random", ["_randommodule.c"]) )
- # high-performance collections
- exts.append( Extension("_collections", ["_collectionsmodule.c"]) )
- # bisect
- exts.append( Extension("_bisect", ["_bisectmodule.c"]) )
- # heapq
- exts.append( Extension("_heapq", ["_heapqmodule.c"]) )
- # operator.add() and similar goodies
- exts.append( Extension('operator', ['operator.c']) )
- # Python 3.0 _fileio module
- exts.append( Extension("_fileio", ["_fileio.c"]) )
- # Python 3.0 _bytesio module
- exts.append( Extension("_bytesio", ["_bytesio.c"]) )
- # _functools
- exts.append( Extension("_functools", ["_functoolsmodule.c"]) )
- # _json speedups
- exts.append( Extension("_json", ["_json.c"]) )
- # Python C API test module
- exts.append( Extension('_testcapi', ['_testcapimodule.c', 'cPickle.c'],
- define_macros=[('NO_STATIC_MEMOTABLE', 1)],
- depends=['testcapi_long.h']) )
- # profilers (_lsprof is for cProfile.py)
- exts.append( Extension('_hotshot', ['_hotshot.c']) )
- exts.append( Extension('_lsprof', ['_lsprof.c', 'rotatingtree.c']) )
- # static Unicode character database
- if have_unicode:
- exts.append( Extension('unicodedata', ['unicodedata.c']) )
- else:
- missing.append('unicodedata')
- # access to ISO C locale support
- data = open('pyconfig.h').read()
- m = re.search(r"#s*define\s+WITH_LIBINTL\s+1\s*", data)
- if m is not None:
- locale_libs = ['intl']
- else:
- locale_libs = []
- if platform == 'darwin':
- locale_extra_link_args = ['-framework', 'CoreFoundation']
- else:
- locale_extra_link_args = []
- exts.append( Extension('_locale', ['_localemodule.c'],
- libraries=locale_libs,
- extra_link_args=locale_extra_link_args) )
- # Modules with some UNIX dependencies -- on by default:
- # (If you have a really backward UNIX, select and socket may not be
- # supported...)
- # fcntl(2) and ioctl(2)
- exts.append( Extension('fcntl', ['fcntlmodule.c']) )
- if platform not in ['mac']:
- # pwd(3)
- exts.append( Extension('pwd', ['pwdmodule.c']) )
- # grp(3)
- exts.append( Extension('grp', ['grpmodule.c']) )
- # spwd, shadow passwords
- if (config_h_vars.get('HAVE_GETSPNAM', False) or
- config_h_vars.get('HAVE_GETSPENT', False)):
- exts.append( Extension('spwd', ['spwdmodule.c']) )
- else:
- missing.append('spwd')
- else:
- missing.extend(['pwd', 'grp', 'spwd'])
- # select(2); not on ancient System V
- exts.append( Extension('select', ['selectmodule.c']) )
- # Fred Drake's interface to the Python parser
- exts.append( Extension('parser', ['parsermodule.c']) )
- # cStringIO and cPickle
- exts.append( Extension('cStringIO', ['cStringIO.c']) )
- exts.append( Extension('cPickle', ['cPickle.c']) )
- # Memory-mapped files (also works on Win32).
- if platform not in ['atheos', 'mac']:
- exts.append( Extension('mmap', ['mmapmodule.c']) )
- else:
- missing.append('mmap')
- # Lance Ellinghaus's syslog module
- if platform not in ['mac']:
- # syslog daemon interface
- exts.append( Extension('syslog', ['syslogmodule.c']) )
- else:
- missing.append('syslog')
- # George Neville-Neil's timing module:
- # Deprecated in PEP 4 http://www.python.org/peps/pep-0004.html
- # http://mail.python.org/pipermail/python-dev/2006-January/060023.html
- #exts.append( Extension('timing', ['timingmodule.c']) )
- #
- # Here ends the simple stuff. From here on, modules need certain
- # libraries, are platform-specific, or present other surprises.
- #
- # Multimedia modules
- # These don't work for 64-bit platforms!!!
- # These represent audio samples or images as strings:
- # Operations on audio samples
- # According to #993173, this one should actually work fine on
- # 64-bit platforms.
- exts.append( Extension('audioop', ['audioop.c']) )
- # Disabled on 64-bit platforms
- if sys.maxint != 9223372036854775807L:
- # Operations on images
- exts.append( Extension('imageop', ['imageop.c']) )
- else:
- missing.extend(['imageop'])
- # readline
- do_readline = self.compiler.find_library_file(lib_dirs, 'readline')
- if platform == 'darwin': # and os.uname()[2] < '9.':
- # MacOSX 10.4 has a broken readline. Don't try to build
- # the readline module unless the user has installed a fixed
- # readline package
- # FIXME: The readline emulation on 10.5 is better, but the
- # readline module doesn't compile out of the box.
- if find_file('readline/rlconf.h', inc_dirs, []) is None:
- do_readline = False
- if do_readline:
- if sys.platform == 'darwin':
- # In every directory on the search path search for a dynamic
- # library and then a static library, instead of first looking
- # for dynamic libraries on the entiry path.
- # This way a staticly linked custom readline gets picked up
- # before the (broken) dynamic library in /usr/lib.
- readline_extra_link_args = ('-Wl,-search_paths_first',)
- else:
- readline_extra_link_args = ()
- readline_libs = ['readline']
- if self.compiler.find_library_file(lib_dirs,
- 'ncursesw'):
- readline_libs.append('ncursesw')
- elif self.compiler.find_library_file(lib_dirs,
- 'ncurses'):
- readline_libs.append('ncurses')
- elif self.compiler.find_library_file(lib_dirs, 'curses'):
- readline_libs.append('curses')
- elif self.compiler.find_library_file(lib_dirs +
- ['/usr/lib/termcap'],
- 'termcap'):
- readline_libs.append('termcap')
- exts.append( Extension('readline', ['readline.c'],
- library_dirs=['/usr/lib/termcap'],
- extra_link_args=readline_extra_link_args,
- libraries=readline_libs) )
- else:
- missing.append('readline')
- if platform not in ['mac']:
- # crypt module.
- if self.compiler.find_library_file(lib_dirs, 'crypt'):
- libs = ['crypt']
- else:
- libs = []
- exts.append( Extension('crypt', ['cryptmodule.c'], libraries=libs) )
- else:
- missing.append('crypt')
- # CSV files
- exts.append( Extension('_csv', ['_csv.c']) )
- # LLVM wrappers
- if sysconfig.get_config_var("WITH_LLVM"):
- exts.append( Extension('_llvm', ['_llvm.c']))
- else:
- missing.append('_llvm')
- # socket(2)
- exts.append( Extension('_socket', ['socketmodule.c'],
- depends = ['socketmodule.h']) )
- # Detect SSL support for the socket module (via _ssl)
- search_for_ssl_incs_in = []
- if use_system_paths:
- search_for_ssl_incs_in = [
- '/usr/local/ssl/include',
- '/usr/contrib/ssl/include/'
- ]
- if "SSL_ROOT" in os.environ:
- search_for_ssl_incs_in.append(os.path.join(os.environ["SSL_ROOT"],
- "include"))
- ssl_incs = find_file('openssl/ssl.h', inc_dirs,
- search_for_ssl_incs_in
- )
- if ssl_incs is not None:
- krb5_h = find_file(
- 'krb5.h', inc_dirs,
- ['/usr/kerberos/include'] if use_system_paths else [])
- if krb5_h:
- ssl_incs += krb5_h
- search_for_ssl_libs_in = []
- if use_system_paths:
- search_for_ssl_libs_in = ['/usr/local/ssl/lib',
- '/usr/contrib/ssl/lib/']
- if "SSL_ROOT" in os.environ:
- search_for_ssl_libs_in.append(os.path.join(os.environ["SSL_ROOT"],
- "lib"))
- ssl_libs = find_library_file(self.compiler, 'ssl',lib_dirs,
- search_for_ssl_libs_in)
- if (ssl_incs is not None and
- ssl_libs is not None):
- exts.append( Extension('_ssl', ['_ssl.c'],
- include_dirs = ssl_incs,
- library_dirs = ssl_libs,
- libraries = ['ssl', 'crypto'],
- depends = ['socketmodule.h']), )
- else:
- missing.append('_ssl')
- # find out which version of OpenSSL we have
- openssl_ver = 0
- openssl_ver_re = re.compile(
- '^\s*#\s*define\s+OPENSSL_VERSION_NUMBER\s+(0x[0-9a-fA-F]+)' )
- for ssl_inc_dir in inc_dirs + search_for_ssl_incs_in:
- name = os.path.join(ssl_inc_dir, 'openssl', 'opensslv.h')
- if os.path.isfile(name):
- try:
- incfile = open(name, 'r')
- for line in incfile:
- m = openssl_ver_re.match(line)
- if m:
- openssl_ver = eval(m.group(1))
- break
- except IOError:
- pass
- # first version found is what we'll use (as the compiler should)
- if openssl_ver:
- break
- #print 'openssl_ver = 0x%08x' % openssl_ver
- if (ssl_incs is not None and
- ssl_libs is not None and
- openssl_ver >= 0x00907000):
- # The _hashlib module wraps optimized implementations
- # of hash functions from the OpenSSL library.
- exts.append( Extension('_hashlib', ['_hashopenssl.c'],
- include_dirs = ssl_incs,
- library_dirs = ssl_libs,
- libraries = ['ssl', 'crypto']) )
- # these aren't strictly missing since they are unneeded.
- #missing.extend(['_sha', '_md5'])
- else:
- # The _sha module implements the SHA1 hash algorithm.
- exts.append( Extension('_sha', ['shamodule.c']) )
- # The _md5 module implements the RSA Data Security, Inc. MD5
- # Message-Digest Algorithm, described in RFC 1321. The
- # necessary files md5.c and md5.h are included here.
- exts.append( Extension('_md5',
- sources = ['md5module.c', 'md5.c'],
- depends = ['md5.h']) )
- missing.append('_hashlib')
- if (openssl_ver < 0x00908000):
- # OpenSSL doesn't do these until 0.9.8 so we'll bring our own hash
- exts.append( Extension('_sha256', ['sha256module.c']) )
- exts.append( Extension('_sha512', ['sha512module.c']) )
- # Modules that provide persistent dictionary-like semantics. You will
- # probably want to arrange for at least one of them to be available on
- # your machine, though none are defined by default because of library
- # dependencies. The Python module anydbm.py provides an
- # implementation independent wrapper for these; dumbdbm.py provides
- # similar functionality (but slower of course) implemented in Python.
- # Sleepycat^WOracle Berkeley DB interface.
- # http://www.oracle.com/database/berkeley-db/db/index.html
- #
- # This requires the Sleepycat^WOracle DB code. The supported versions
- # are set below. Visit the URL above to download
- # a release. Most open source OSes come with one or more
- # versions of BerkeleyDB already installed.
- max_db_ver = (4, 7)
- min_db_ver = (3, 3)
- db_setup_debug = False # verbose debug prints from this script?
- def allow_db_ver(db_ver):
- """Returns a boolean if the given BerkeleyDB version is acceptable.
- Args:
- db_ver: A tuple of the version to verify.
- """
- if not (min_db_ver <= db_ver <= max_db_ver):
- return False
- # Use this function to filter out known bad configurations.
- if (4, 6) == db_ver[:2]:
- # BerkeleyDB 4.6.x is not stable on many architectures.
- arch = platform_machine()
- if arch not in ('i386', 'i486', 'i586', 'i686',
- 'x86_64', 'ia64'):
- return False
- return True
- def gen_db_minor_ver_nums(major):
- if major == 4:
- for x in range(max_db_ver[1]+1):
- if allow_db_ver((4, x)):
- yield x
- elif major == 3:
- for x in (3,):
- if allow_db_ver((3, x)):
- yield x
- else:
- raise ValueError("unknown major BerkeleyDB version", major)
- # construct a list of paths to look for the header file in on
- # top of the normal inc_dirs.
- db_inc_paths = []
- if use_system_paths:
- db_inc_paths.extend([
- '/usr/include/db4',
- '/usr/local/include/db4',
- '/opt/sfw/include/db4',
- '/usr/include/db3',
- '/usr/local/include/db3',
- '/opt/sfw/include/db3',
- # Fink defaults (http://fink.sourceforge.net/)
- '/sw/include/db4',
- '/sw/include/db3',
- ])
- # 4.x minor number specific paths
- for x in gen_db_minor_ver_nums(4):
- db_inc_paths.append('/usr/include/db4%d' % x)
- db_inc_paths.append('/usr/include/db4.%d' % x)
- db_inc_paths.append('/usr/local/BerkeleyDB.4.%d/include' % x)
- db_inc_paths.append('/usr/local/include/db4%d' % x)
- db_inc_paths.append('/pkg/db-4.%d/include' % x)
- db_inc_paths.append('/opt/db-4.%d/include' % x)
- # MacPorts default (http://www.macports.org/)
- db_inc_paths.append('/opt/local/include/db4%d' % x)
- # 3.x minor number specific paths
- for x in gen_db_minor_ver_nums(3):
- db_inc_paths.append('/usr/include/db3%d' % x)
- db_inc_paths.append('/usr/local/BerkeleyDB.3.%d/include' % x)
- db_inc_paths.append('/usr/local/include/db3%d' % x)
- db_inc_paths.append('/pkg/db-3.%d/include' % x)
- db_inc_paths.append('/opt/db-3.%d/include' % x)
- # Add some common subdirectories for Sleepycat DB to the list,
- # based on the standard include directories. This way DB3/4 gets
- # picked up when it is installed in a non-standard prefix and
- # the user has added that prefix into inc_dirs.
- std_variants = []
- for dn in inc_dirs:
- std_variants.append(os.path.join(dn, 'db3'))
- std_variants.append(os.path.join(dn, 'db4'))
- for x in gen_db_minor_ver_nums(4):
- std_variants.append(os.path.join(dn, "db4%d"%x))
- std_variants.append(os.path.join(dn, "db4.%d"%x))
- for x in gen_db_minor_ver_nums(3):
- std_variants.append(os.path.join(dn, "db3%d"%x))
- std_variants.append(os.path.join(dn, "db3.%d"%x))
- db_inc_paths = std_variants + db_inc_paths
- db_inc_paths = [p for p in db_inc_paths if os.path.exists(p)]
- db_ver_inc_map = {}
- class db_found(Exception): pass
- try:
- # See whether there is a Sleepycat header in the standard
- # search path.
- for d in inc_dirs + db_inc_paths:
- f = os.path.join(d, "db.h")
- if db_setup_debug: print "db: looking for db.h in", f
- if os.path.exists(f):
- f = open(f).read()
- m = re.search(r"#define\WDB_VERSION_MAJOR\W(\d+)", f)
- if m:
- db_major = int(m.group(1))
- m = re.search(r"#define\WDB_VERSION_MINOR\W(\d+)", f)
- db_minor = int(m.group(1))
- db_ver = (db_major, db_minor)
- # Avoid 4.6 prior to 4.6.21 due to a BerkeleyDB bug
- if db_ver == (4, 6):
- m = re.search(r"#define\WDB_VERSION_PATCH\W(\d+)", f)
- db_patch = int(m.group(1))
- if db_patch < 21:
- print "db.h:", db_ver, "patch", db_patch,
- print "being ignored (4.6.x must be >= 4.6.21)"
- continue
- if ( (not db_ver_inc_map.has_key(db_ver)) and
- allow_db_ver(db_ver) ):
- # save the include directory with the db.h version
- # (first occurrence only)
- db_ver_inc_map[db_ver] = d
- if db_setup_debug:
- print "db.h: found", db_ver, "in", d
- else:
- # we already found a header for this library version
- if db_setup_debug: print "db.h: ignoring", d
- else:
- # ignore this header, it didn't contain a version number
- if db_setup_debug:
- print "db.h: no version number version in", d
- db_found_vers = db_ver_inc_map.keys()
- db_found_vers.sort()
- while db_found_vers:
- db_ver = db_found_vers.pop()
- db_incdir = db_ver_inc_map[db_ver]
- # check lib directories parallel to the location of the header
- db_dirs_to_check = combine_dirs_to_check([
- db_incdir.replace("include", 'lib64'),
- db_incdir.replace("include", 'lib'),
- ], lib_dirs)
- # Look for a version specific db-X.Y before an ambiguoius dbX
- # XXX should we -ever- look for a dbX name? Do any
- # systems really not name their library by version and
- # symlink to more general names?
- for dblib in (('db-%d.%d' % db_ver),
- ('db%d%d' % db_ver),
- ('db%d' % db_ver[0])):
- dblib_file = self.compiler.find_library_file(
- db_dirs_to_check, dblib )
- if dblib_file:
- dblib_dir = [ os.path.abspath(os.path.dirname(dblib_file)) ]
- raise db_found
- else:
- if db_setup_debug: print "db lib: ", dblib, "not found"
- except db_found:
- if db_setup_debug:
- print "bsddb using BerkeleyDB lib:", db_ver, dblib
- print "bsddb lib dir:", dblib_dir, " inc dir:", db_incdir
- db_incs = [db_incdir]
- dblibs = [dblib]
- # We add the runtime_library_dirs argument because the
- # BerkeleyDB lib we're linking against often isn't in the
- # system dynamic library search path. This is usually
- # correct and most trouble free, but may cause problems in
- # some unusual system configurations (e.g. the directory
- # is on an NFS server that goes away).
- exts.append(Extension('_bsddb', ['_bsddb.c'],
- depends = ['bsddb.h'],
- library_dirs=dblib_dir,
- runtime_library_dirs=dblib_dir,
- include_dirs=db_incs,
- libraries=dblibs))
- else:
- if db_setup_debug: print "db: no appropriate library found"
- db_incs = None
- dblibs = []
- dblib_dir = None
- missing.append('_bsddb')
- # The sqlite interface
- sqlite_setup_debug = False # verbose debug prints from this script?
- # We hunt for #define SQLITE_VERSION "n.n.n"
- # We need to find >= sqlite version 3.0.8
- sqlite_incdir = sqlite_libdir = None
- sqlite_inc_paths = []
- if use_system_paths:
- sqlite_inc_paths = [ '/usr/include',
- '/usr/include/sqlite',
- '/usr/include/sqlite3',
- '/usr/local/include',
- '/usr/local/include/sqlite',
- '/usr/local/include/sqlite3',
- ]
- if "SQLITE_INC" in os.environ:
- sqlite_inc_paths.append(os.environ["SQLITE_INC"])
- MIN_SQLITE_VERSION_NUMBER = (3, 0, 8)
- MIN_SQLITE_VERSION = ".".join([str(x)
- for x in MIN_SQLITE_VERSION_NUMBER])
- # Scan the default include directories before the SQLite specific
- # ones. This allows one to override the copy of sqlite on OSX,
- # where /usr/include contains an old version of sqlite.
- for d in inc_dirs + sqlite_inc_paths:
- f = os.path.join(d, "sqlite3.h")
- if sqlite_setup_debug: print "sqlite: looking for sqlite3.h in", f
- if os.path.exists(f):
- if sqlite_setup_debug: print "sqlite: found %s"%f
- incf = open(f).read()
- m = re.search(
- r'\s*.*#\s*.*define\s.*SQLITE_VERSION\W*"(.*)"', incf)
- if m:
- sqlite_version = m.group(1)
- sqlite_version_tuple = tuple([int(x)
- for x in sqlite_version.split(".")])
- if sqlite_version_tuple >= MIN_SQLITE_VERSION_NUMBER:
- # we win!
- if sqlite_setup_debug:
- print "%s/sqlite3.h: version %s"%(d, sqlite_version)
- sqlite_incdir = d
- break
- else:
- if sqlite_setup_debug:
- print "%s: version %d is too old, need >= %s"%(d,
- sqlite_version, MIN_SQLITE_VERSION)
- elif sqlite_setup_debug:
- print "sqlite: %s had no SQLITE_VERSION"%(f,)
- if sqlite_incdir:
- sqlite_dirs_to_check = combine_dirs_to_check([
- os.path.join(sqlite_incdir, '..', 'lib64'),
- os.path.join(sqlite_incdir, '..', 'lib'),
- os.path.join(sqlite_incdir, '..', '..', 'lib64'),
- os.path.join(sqlite_incdir, '..', '..', 'lib'),
- os.environ.get("SQLITE_LIB", ""),
- ], lib_dirs)
- sqlite_libfile = self.compiler.find_library_file(
- sqlite_dirs_to_check, 'sqlite3')
- if sqlite_libfile:
- sqlite_libdir = [os.path.abspath(os.path.dirname(sqlite_libfile))]
- elif sqlite_setup_debug:
- print ("sqlite: no sqlite3 library in %s" %
- (sqlite_dirs_to_check))
- if sqlite_incdir and sqlite_libdir:
- sqlite_srcs = ['_sqlite/cache.c',
- '_sqlite/connection.c',
- '_sqlite/cursor.c',
- '_sqlite/microprotocols.c',
- '_sqlite/module.c',
- '_sqlite/prepare_protocol.c',
- '_sqlite/row.c',
- '_sqlite/statement.c',
- '_sqlite/util.c', ]
- sqlite_defines = []
- if sys.platform != "win32":
- sqlite_defines.append(('MODULE_NAME', '"sqlite3"'))
- else:
- sqlite_defines.append(('MODULE_NAME', '\\"sqlite3\\"'))
- if sys.platform == 'darwin':
- # In every directory on the search path search for a dynamic
- # library and then a static library, instead of first looking
- # for dynamic libraries on the entiry path.
- # This way a staticly linked custom sqlite gets picked up
- # before the dynamic library in /usr/lib.
- sqlite_extra_link_args = ('-Wl,-search_paths_first',)
- else:
- sqlite_extra_link_args = ()
- exts.append(Extension('_sqlite3', sqlite_srcs,
- define_macros=sqlite_defines,
- include_dirs=["Modules/_sqlite",
- sqlite_incdir],
- library_dirs=sqlite_libdir,
- runtime_library_dirs=sqlite_libdir,
- extra_link_args=sqlite_extra_link_args,
- libraries=["sqlite3",]))
- else:
- missing.append('_sqlite3')
- # Look for Berkeley db 1.85. Note that it is built as a different
- # module name so it can be included even when later versions are
- # available. A very restrictive search is performed to avoid
- # accidentally building this module with a later version of the
- # underlying db library. May BSD-ish Unixes incorporate db 1.85
- # symbols into libc and place the include file in /usr/include.
- #
- # If the better bsddb library can be built (db_incs is defined)
- # we do not build this one. Otherwise this build will pick up
- # the more recent berkeleydb's db.h file first in the include path
- # when attempting to compile and it will fail.
- f = "/usr/include/db.h"
- if use_system_paths and os.path.exists(f) and not db_incs:
- data = open(f).read()
- m = re.search(r"#s*define\s+HASHVERSION\s+2\s*", data)
- if m is not None:
- # bingo - old version used hash file format version 2
- ### XXX this should be fixed to not be platform-dependent
- ### but I don't have direct access to an osf1 platform and
- ### seemed to be muffing the search somehow
- libraries = platform == "osf1" and ['db'] or None
- if libraries is not None:
- exts.append(Extension('bsddb185', ['bsddbmodule.c'],
- libraries=libraries))
- else:
- exts.append(Extension('bsddb185', ['bsddbmodule.c']))
- else:
- missing.append('bsddb185')
- else:
- missing.append('bsddb185')
- # The standard Unix dbm module:
- if platform not in ['cygwin']:
- if find_file("ndbm.h", inc_dirs, []) is not None:
- # Some systems have -lndbm, others don't
- if self.compiler.find_library_file(lib_dirs, 'ndbm'):
- ndbm_libs = ['ndbm']
- else:
- ndbm_libs = []
- exts.append( Extension('dbm', ['dbmmodule.c'],
- define_macros=[('HAVE_NDBM_H',None)],
- libraries = ndbm_libs ) )
- elif self.compiler.find_library_file(lib_dirs, 'gdbm'):
- gdbm_libs = ['gdbm']
- if self.compiler.find_library_file(lib_dirs, 'gdbm_compat'):
- gdbm_libs.append('gdbm_compat')
- if find_file("gdbm/ndbm.h", inc_dirs, []) is not None:
- exts.append( Extension(
- 'dbm', ['dbmmodule.c'],
- define_macros=[('HAVE_GDBM_NDBM_H',None)],
- libraries = gdbm_libs ) )
- elif find_file("gdbm-ndbm.h", inc_dirs, []) is not None:
- exts.append( Extension(
- 'dbm', ['dbmmodule.c'],
- define_macros=[('HAVE_GDBM_DASH_NDBM_H',None)],
- libraries = gdbm_libs ) )
- else:
- missing.append('dbm')
- elif db_incs is not None:
- exts.append( Extension('dbm', ['dbmmodule.c'],
- library_dirs=dblib_dir,
- runtime_library_dirs=dblib_dir,
- include_dirs=db_incs,
- define_macros=[('HAVE_BERKDB_H',None),
- ('DB_DBM_HSEARCH',None)],
- libraries=dblibs))
- else:
- missing.append('dbm')
- # Anthony Baxter's gdbm module. GNU dbm(3) will require -lgdbm:
- if (self.compiler.find_library_file(lib_dirs, 'gdbm')):
- exts.append( Extension('gdbm', ['gdbmmodule.c'],
- libraries = ['gdbm'] ) )
- else:
- missing.append('gdbm')
- # Unix-only modules
- if platform not in ['mac', 'win32']:
- # Steen Lumholt's termios module
- exts.append( Extension('termios', ['termios.c']) )
- # Jeremy Hylton's rlimit interface
- if platform not in ['atheos']:
- exts.append( Extension('resource', ['resource.c']) )
- else:
- missing.append('resource')
- # Sun yellow pages. Some systems have the functions in libc.
- if platform not in ['cygwin', 'atheos', 'qnx6']:
- if (self.compiler.find_libra…