components/python/python26/Python26-09-ucred.patch
changeset 99 c15c9099bb44
--- /dev/null	Thu Jan 01 00:00:00 1970 +0000
+++ b/components/python/python26/Python26-09-ucred.patch	Wed Feb 23 10:37:11 2011 -0800
@@ -0,0 +1,2427 @@
+diff --git Python-2.6.4/Modules/ucred.c Python-2.6.4/Modules/ucred.c
+new file mode 100644
+--- /dev/null
++++ Python-2.6.4/Modules/ucred.c
+@@ -0,0 +1,391 @@
++/*
++ * Permission is hereby granted, free of charge, to any person obtaining a copy
++ * of this software and associated documentation files (the "Software"), to
++ * deal in the Software without restriction, including without limitation the
++ * rights to use, copy, modify, merge, publish, distribute, sublicense, and/or
++ * sell copies of the Software, and to permit persons to whom the Software is
++ * furnished to do so, subject to the following conditions:
++ *
++ * The above copyright notice and this permission notice shall be included in
++ * all copies or substantial portions of the Software.
++ *
++ * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
++ * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
++ * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE
++ * AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
++ * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
++ * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER
++ * DEALINGS IN THE SOFTWARE.
++ *
++ * Copyright 2008 Sun Microsystems, Inc.  All rights reserved.
++ * Use is subject to license terms.
++ */
++
++#include <Python.h>
++
++#include <stdio.h>
++#include <priv.h>
++#include <ucred.h>
++#include <ctype.h>
++#include <tsol/label.h>
++
++typedef struct {
++	PyObject_HEAD
++	ucred_t *ucred;
++} pyucred_t;
++
++#define pyucred_getlongid(name, type)				\
++	static PyObject *					\
++	pyucred_get##name(pyucred_t *uc)			\
++	{ 							\
++		type val;					\
++								\
++		if (uc->ucred == NULL) {			\
++			errno = EINVAL;				\
++			PyErr_SetFromErrno(PyExc_OSError);	\
++			return (NULL);				\
++		}						\
++								\
++		if ((val = ucred_get##name(uc->ucred)) == -1) {	\
++			PyErr_SetFromErrno(PyExc_OSError);	\
++			return (NULL);				\
++		}						\
++								\
++		return (Py_BuildValue("l", (long)val));		\
++	}
++
++pyucred_getlongid(euid, uid_t)
++pyucred_getlongid(ruid, uid_t)
++pyucred_getlongid(suid, uid_t)
++pyucred_getlongid(egid, gid_t)
++pyucred_getlongid(rgid, gid_t)
++pyucred_getlongid(sgid, gid_t)
++pyucred_getlongid(pid, pid_t)
++pyucred_getlongid(projid, projid_t)
++pyucred_getlongid(zoneid, zoneid_t)
++
++static PyObject *
++pyucred_getgroups(pyucred_t *uc)
++{
++	const gid_t *groups;
++	PyObject *list;
++	int len;
++	int i;
++
++	if (uc->ucred == NULL) {
++		errno = EINVAL;
++		PyErr_SetFromErrno(PyExc_OSError);
++		return (NULL);
++	}
++
++	if ((len = ucred_getgroups(uc->ucred, &groups)) == -1) {
++		PyErr_SetFromErrno(PyExc_OSError);
++		return (NULL);
++	}
++
++	if ((list = PyList_New(len)) == NULL)
++		return (NULL);
++
++	for (i = 0; i < len; i++) {
++		PyObject *gid = Py_BuildValue("l", (long)groups[i]);
++		if (PyList_SetItem(list, i, gid) == -1)
++			return (NULL);
++	}
++
++	return (list);
++}
++
++static PyObject *
++pyucred_getlabel(pyucred_t *uc)
++{
++	m_label_t *label;
++	PyObject *ret;
++	char *str;
++
++	if (uc->ucred == NULL) {
++		errno = EINVAL;
++		PyErr_SetFromErrno(PyExc_OSError);
++		return (NULL);
++	}
++
++	label = ucred_getlabel(uc->ucred);
++	if (label == NULL)
++		return (Py_BuildValue("s", ""));
++
++	if (label_to_str(label, &str, M_LABEL, DEF_NAMES) == -1) {
++		PyErr_SetFromErrno(PyExc_OSError);
++		return (NULL);
++	}
++
++	ret = Py_BuildValue("s", str);
++	free(str);
++	return (ret);
++}
++
++static PyObject *
++pyucred_getpflags(pyucred_t *uc, PyObject *args, PyObject *kwargs)
++{
++	static char *kwlist[] = { "flags", NULL };
++	uint_t flags;
++
++	if (uc->ucred == NULL) {
++		errno = EINVAL;
++		PyErr_SetFromErrno(PyExc_OSError);
++		return (NULL);
++	}
++
++	if (!PyArg_ParseTupleAndKeywords(args, kwargs, "i", kwlist,
++	    &flags))
++		return (NULL);
++
++	if ((flags = ucred_getpflags(uc->ucred, flags)) == (uint_t)-1) {
++		PyErr_SetFromErrno(PyExc_OSError);
++		return (NULL);
++	}
++
++	return (Py_BuildValue("i", flags));
++}
++
++static PyObject *
++pyucred_has_priv(pyucred_t *uc, PyObject *args, PyObject *kwargs)
++{
++	static char *kwlist[] = { "set", "priv", NULL };
++	const priv_set_t *privs;
++	const char *set;
++	const char *priv;
++
++	if (uc->ucred == NULL) {
++		errno = EINVAL;
++		PyErr_SetFromErrno(PyExc_OSError);
++		return (NULL);
++	}
++
++	if (!PyArg_ParseTupleAndKeywords(args, kwargs, "ss", kwlist,
++	    &set, &priv))
++		return (NULL);
++
++	if ((privs = ucred_getprivset(uc->ucred, set)) == NULL) {
++		PyErr_SetFromErrno(PyExc_OSError);
++		return (NULL);
++	}
++
++	if (priv_ismember(privs, priv)) {
++		Py_INCREF(Py_True);
++		return Py_True;
++	}
++
++	Py_INCREF(Py_False);
++	return Py_False;
++}
++
++PyDoc_STRVAR(pyucred_getlabel_doc,
++    "getlabel() -> string\n"
++    "\n"
++    "Return the Trusted Extensions label string, or an "
++    "empty string if not available. The label string is "
++    "converted using the default name and M_LABEL (human-readable). "
++    "Raises OSError. See label_to_str(3TSOL).");
++PyDoc_STRVAR(pyucred_getpflags_doc,
++    "getpflags(flags) -> int\n"
++    "\n"
++    "Return the values of the specified privilege flags.");
++PyDoc_STRVAR(pyucred_has_priv_doc,
++    "has_priv(set, priv) -> bool\n"
++    "\n"
++    "Return true if the given privilege is set in the "
++    "specified set. Raises OSError if the set or privilege is "
++    "invalid, or a problem occurs.\n"
++    "\n"
++    "Currently, the following privilege sets are defined, as "
++    "described in privileges(5):\n"
++    "\n"
++    "Effective\n"
++    "Permitted\n"
++    "Inheritable\n"
++    "Limit\n");
++
++static PyMethodDef pyucred_methods[] = {
++	{ "geteuid", (PyCFunction)pyucred_geteuid, METH_NOARGS,
++	    "Return the effective user ID." },
++	{ "getruid", (PyCFunction)pyucred_getruid, METH_NOARGS,
++	    "Return the real user ID." },
++	{ "getsuid", (PyCFunction)pyucred_getsuid, METH_NOARGS,
++	    "Return the saved user ID." },
++	{ "getegid", (PyCFunction)pyucred_getegid, METH_NOARGS,
++	    "Return the effective group ID." },
++	{ "getrgid", (PyCFunction)pyucred_getrgid, METH_NOARGS,
++	    "Return the real group ID." },
++	{ "getsgid", (PyCFunction)pyucred_getsgid, METH_NOARGS,
++	    "Return the saved group ID." },
++	{ "getpid", (PyCFunction)pyucred_getpid, METH_NOARGS,
++	    "Return the effective user ID." },
++	{ "getprojid", (PyCFunction)pyucred_getprojid, METH_NOARGS,
++	    "Return the project ID." },
++	{ "getzoneid", (PyCFunction)pyucred_getzoneid, METH_NOARGS,
++	    "Return the zone ID." },
++	{ "getgroups", (PyCFunction)pyucred_getgroups, METH_NOARGS,
++	    "Return a list of group IDs." },
++	{ "getlabel", (PyCFunction)pyucred_getlabel, METH_NOARGS,
++	    pyucred_getlabel_doc },
++	{ "getpflags", (PyCFunction)pyucred_getpflags,
++	    METH_VARARGS|METH_KEYWORDS, pyucred_getpflags_doc },
++	{ "has_priv", (PyCFunction)pyucred_has_priv,
++	    METH_VARARGS|METH_KEYWORDS, pyucred_has_priv_doc },
++	{ NULL }
++};
++
++static int
++pyucred_init(PyObject *self, PyObject *args, PyObject *kwargs)
++{
++	pyucred_t *uc = (pyucred_t *)self;
++	uc->ucred = NULL;
++	return (0);
++}
++
++static void
++pyucred_dealloc(PyObject *self)
++{
++	pyucred_t *uc = (pyucred_t *)self;
++	if (uc->ucred != NULL)
++		ucred_free(uc->ucred);
++	self->ob_type->tp_free(self);
++}
++
++static PyTypeObject pyucred_type = {
++	PyObject_HEAD_INIT(NULL)
++	0,                         /*ob_size*/
++	"ucred.ucred",             /*tp_name*/
++	sizeof (pyucred_t),        /*tp_basicsize*/
++	0,                         /*tp_itemsize*/
++	pyucred_dealloc,           /*tp_dealloc*/
++	0,                         /*tp_print*/
++	0,                         /*tp_getattr*/
++	0,                         /*tp_setattr*/
++	0,                         /*tp_compare*/
++	0,                         /*tp_repr*/
++	0,                         /*tp_as_number*/
++	0,                         /*tp_as_sequence*/
++	0,                         /*tp_as_mapping*/
++	0,                         /*tp_hash */
++	0,                         /*tp_call*/
++	0,                         /*tp_str*/
++	0,                         /*tp_getattro*/
++	0,                         /*tp_setattro*/
++	0,                         /*tp_as_buffer*/
++	Py_TPFLAGS_DEFAULT,        /*tp_flags*/
++	"user credentials",        /*tp_doc */
++	0,		           /* tp_traverse */
++	0,		           /* tp_clear */
++	0,		           /* tp_richcompare */
++	0,		           /* tp_weaklistoffset */
++	0,		           /* tp_iter */
++	0,		           /* tp_iternext */
++	pyucred_methods,           /* tp_methods */
++	0,                         /* tp_members */
++	0,                         /* tp_getset */
++	0,                         /* tp_base */
++	0,                         /* tp_dict */
++	0,                         /* tp_descr_get */
++	0,                         /* tp_descr_set */
++	0,                         /* tp_dictoffset */
++	(initproc)pyucred_init,    /* tp_init */
++	0,                         /* tp_alloc */
++	0,                         /* tp_new */
++};
++
++static PyObject *
++pyucred_new(const ucred_t *uc)
++{
++	pyucred_t *self;
++
++	self = (pyucred_t *)PyObject_CallObject((PyObject *)&pyucred_type, NULL);
++
++	if (self == NULL)
++		return (NULL);
++
++	self->ucred = (ucred_t *)uc;
++
++	return ((PyObject *)self);
++}
++
++static PyObject *
++pyucred_get(PyObject *o, PyObject *args, PyObject *kwargs)
++{
++	static char *kwlist[] = { "pid", NULL };
++	ucred_t *ucred = NULL;
++	int pid;
++
++	if (!PyArg_ParseTupleAndKeywords(args, kwargs, "i", kwlist,
++	    &pid))
++		return (NULL);
++
++	ucred = ucred_get(pid);
++
++	if (ucred == NULL) {
++		PyErr_SetFromErrno(PyExc_OSError);
++		return (NULL);
++	}
++
++	return (pyucred_new(ucred));
++}
++
++static PyObject *
++pyucred_getpeer(PyObject *o, PyObject *args, PyObject *kwargs)
++{
++	static char *kwlist[] = { "fd", NULL };
++	ucred_t *ucred = NULL;
++	int fd;
++
++	if (!PyArg_ParseTupleAndKeywords(args, kwargs, "i", kwlist,
++	    &fd))
++		return (NULL);
++
++	if (getpeerucred(fd, &ucred) == -1) {
++		PyErr_SetFromErrno(PyExc_OSError);
++		return (NULL);
++	}
++
++	return (pyucred_new(ucred));
++}
++
++PyDoc_STRVAR(pyucred_get_doc,
++    "get(pid) -> ucred\n"
++    "\n"
++    "Return the credentials of the specified process ID. "
++    "Raises OSError. See ucred_get(3C).");
++PyDoc_STRVAR(pyucred_getpeer_doc,
++    "getpeer(fd) -> ucred\n"
++    "\n"
++    "Return the credentials of the peer endpoint of a "
++    "connection-oriented socket (SOCK_STREAM) or STREAM fd "
++    "at the time the endpoint was created or the connection "
++    "was established. Raises OSError. See getpeerucred(3C).");
++
++static struct PyMethodDef pyucred_module_methods[] = {
++	{ "get", (PyCFunction) pyucred_get,
++	  METH_VARARGS|METH_KEYWORDS, pyucred_get_doc },
++	{ "getpeer", (PyCFunction) pyucred_getpeer,
++	  METH_VARARGS|METH_KEYWORDS, pyucred_getpeer_doc },
++	{ NULL, NULL, 0, NULL }
++};
++
++PyDoc_STRVAR(pyucred_module_doc,
++    "This module provides an interface to the user credential access "
++    "methods, obtainable either by process ID or file descriptor.");
++   
++PyMODINIT_FUNC
++initucred(void)
++{
++	PyObject *m;
++
++	m = Py_InitModule3("ucred", pyucred_module_methods,
++	    pyucred_module_doc);
++
++	pyucred_type.tp_new = PyType_GenericNew;
++	if (PyType_Ready(&pyucred_type) < 0)
++		return;
++
++	Py_INCREF(&pyucred_type);
++
++	PyModule_AddObject(m, "ucred", (PyObject *)&pyucred_type);
++}
+diff --git Python-2.6.4/setup.py Python-2.6.4/setup.py
+--- Python-2.6.4/setup.py
++++ Python-2.6.4/setup.py
+@@ -1277,6 +1277,13 @@
+         else:
+             missing.append('dl')
+ 
++        # ucred module (Solaris)
++        ucred_inc = find_file('ucred.h', [], inc_dirs)
++        tsol_inc = find_file('tsol/label.h', [], inc_dirs)
++        if ucred_inc is not None and tsol_inc is not None:
++            exts.append( Extension('ucred', ['ucred.c'],
++                                   libraries = ['tsol']) )
++
+         # Thomas Heller's _ctypes module
+         self.detect_ctypes(inc_dirs, lib_dirs)
+ 
+diff --git Python-2.6.4/setup.py.orig Python-2.6.4/setup.py.orig
+new file mode 100644
+--- /dev/null
++++ Python-2.6.4/setup.py.orig
+@@ -0,0 +1,1958 @@
++# Autodetecting setup.py script for building the Python extensions
++#
++
++__version__ = "$Revision: 75282 $"
++
++import sys, os, imp, re, optparse
++from glob import glob
++from platform import machine as platform_machine
++
++from distutils import log
++from distutils import sysconfig
++from distutils import text_file
++from distutils.errors import *
++from distutils.core import Extension, setup
++from distutils.command.build_ext import build_ext
++from distutils.command.install import install
++from distutils.command.install_lib import install_lib
++
++# This global variable is used to hold the list of modules to be disabled.
++disabled_module_list = []
++
++def add_dir_to_list(dirlist, dir):
++    """Add the directory 'dir' to the list 'dirlist' (at the front) if
++    1) 'dir' is not already in 'dirlist'
++    2) 'dir' actually exists, and is a directory."""
++    if dir is not None and os.path.isdir(dir) and dir not in dirlist:
++        dirlist.insert(0, dir)
++
++def find_file(filename, std_dirs, paths):
++    """Searches for the directory where a given file is located,
++    and returns a possibly-empty list of additional directories, or None
++    if the file couldn't be found at all.
++
++    'filename' is the name of a file, such as readline.h or libcrypto.a.
++    'std_dirs' is the list of standard system directories; if the
++        file is found in one of them, no additional directives are needed.
++    'paths' is a list of additional locations to check; if the file is
++        found in one of them, the resulting list will contain the directory.
++    """
++
++    # Check the standard locations
++    for dir in std_dirs:
++        f = os.path.join(dir, filename)
++        if os.path.exists(f): return []
++
++    # Check the additional directories
++    for dir in paths:
++        f = os.path.join(dir, filename)
++        if os.path.exists(f):
++            return [dir]
++
++    # Not found anywhere
++    return None
++
++def find_library_file(compiler, libname, std_dirs, paths):
++    result = compiler.find_library_file(std_dirs + paths, libname)
++    if result is None:
++        return None
++
++    # Check whether the found file is in one of the standard directories
++    dirname = os.path.dirname(result)
++    for p in std_dirs:
++        # Ensure path doesn't end with path separator
++        p = p.rstrip(os.sep)
++        if p == dirname:
++            return [ ]
++
++    # Otherwise, it must have been in one of the additional directories,
++    # so we have to figure out which one.
++    for p in paths:
++        # Ensure path doesn't end with path separator
++        p = p.rstrip(os.sep)
++        if p == dirname:
++            return [p]
++    else:
++        assert False, "Internal error: Path not found in std_dirs or paths"
++
++def module_enabled(extlist, modname):
++    """Returns whether the module 'modname' is present in the list
++    of extensions 'extlist'."""
++    extlist = [ext for ext in extlist if ext.name == modname]
++    return len(extlist)
++
++def find_module_file(module, dirlist):
++    """Find a module in a set of possible folders. If it is not found
++    return the unadorned filename"""
++    list = find_file(module, [], dirlist)
++    if not list:
++        return module
++    if len(list) > 1:
++        log.info("WARNING: multiple copies of %s found"%module)
++    return os.path.join(list[0], module)
++
++class PyBuildExt(build_ext):
++
++    def __init__(self, dist):
++        build_ext.__init__(self, dist)
++        self.failed = []
++
++    def build_extensions(self):
++
++        # Detect which modules should be compiled
++        missing = self.detect_modules()
++
++        # Remove modules that are present on the disabled list
++        extensions = [ext for ext in self.extensions
++                      if ext.name not in disabled_module_list]
++        # move ctypes to the end, it depends on other modules
++        ext_map = dict((ext.name, i) for i, ext in enumerate(extensions))
++        if "_ctypes" in ext_map:
++            ctypes = extensions.pop(ext_map["_ctypes"])
++            extensions.append(ctypes)
++        self.extensions = extensions
++
++        # Fix up the autodetected modules, prefixing all the source files
++        # with Modules/ and adding Python's include directory to the path.
++        (srcdir,) = sysconfig.get_config_vars('srcdir')
++        if not srcdir:
++            # Maybe running on Windows but not using CYGWIN?
++            raise ValueError("No source directory; cannot proceed.")
++
++        # Figure out the location of the source code for extension modules
++        # (This logic is copied in distutils.test.test_sysconfig,
++        # so building in a separate directory does not break test_distutils.)
++        moddir = os.path.join(os.getcwd(), srcdir, 'Modules')
++        moddir = os.path.normpath(moddir)
++        srcdir, tail = os.path.split(moddir)
++        srcdir = os.path.normpath(srcdir)
++        moddir = os.path.normpath(moddir)
++
++        moddirlist = [moddir]
++        incdirlist = ['./Include']
++
++        # Platform-dependent module source and include directories
++        platform = self.get_platform()
++        if platform in ('darwin', 'mac') and ("--disable-toolbox-glue" not in
++            sysconfig.get_config_var("CONFIG_ARGS")):
++            # Mac OS X also includes some mac-specific modules
++            macmoddir = os.path.join(os.getcwd(), srcdir, 'Mac/Modules')
++            moddirlist.append(macmoddir)
++            incdirlist.append('./Mac/Include')
++
++        alldirlist = moddirlist + incdirlist
++
++        # Fix up the paths for scripts, too
++        self.distribution.scripts = [os.path.join(srcdir, filename)
++                                     for filename in self.distribution.scripts]
++
++        # Python header files
++        headers = glob("Include/*.h") + ["pyconfig.h"]
++
++        for ext in self.extensions[:]:
++            ext.sources = [ find_module_file(filename, moddirlist)
++                            for filename in ext.sources ]
++            if ext.depends is not None:
++                ext.depends = [find_module_file(filename, alldirlist)
++                               for filename in ext.depends]
++            else:
++                ext.depends = []
++            # re-compile extensions if a header file has been changed
++            ext.depends.extend(headers)
++
++            ext.include_dirs.append( '.' ) # to get config.h
++            for incdir in incdirlist:
++                ext.include_dirs.append( os.path.join(srcdir, incdir) )
++
++            # If a module has already been built statically,
++            # don't build it here
++            if ext.name in sys.builtin_module_names:
++                self.extensions.remove(ext)
++
++        if platform != 'mac':
++            # Parse Modules/Setup and Modules/Setup.local to figure out which
++            # modules are turned on in the file.
++            remove_modules = []
++            for filename in ('Modules/Setup', 'Modules/Setup.local'):
++                input = text_file.TextFile(filename, join_lines=1)
++                while 1:
++                    line = input.readline()
++                    if not line: break
++                    line = line.split()
++                    remove_modules.append(line[0])
++                input.close()
++
++            for ext in self.extensions[:]:
++                if ext.name in remove_modules:
++                    self.extensions.remove(ext)
++
++        # When you run "make CC=altcc" or something similar, you really want
++        # those environment variables passed into the setup.py phase.  Here's
++        # a small set of useful ones.
++        compiler = os.environ.get('CC')
++        args = {}
++        # unfortunately, distutils doesn't let us provide separate C and C++
++        # compilers
++        if compiler is not None:
++            (ccshared,cflags) = sysconfig.get_config_vars('CCSHARED','CFLAGS')
++            args['compiler_so'] = compiler + ' ' + ccshared + ' ' + cflags
++        self.compiler.set_executables(**args)
++
++        build_ext.build_extensions(self)
++
++        longest = max([len(e.name) for e in self.extensions])
++        if self.failed:
++            longest = max(longest, max([len(name) for name in self.failed]))
++
++        def print_three_column(lst):
++            lst.sort(key=str.lower)
++            # guarantee zip() doesn't drop anything
++            while len(lst) % 3:
++                lst.append("")
++            for e, f, g in zip(lst[::3], lst[1::3], lst[2::3]):
++                print "%-*s   %-*s   %-*s" % (longest, e, longest, f,
++                                              longest, g)
++
++        if missing:
++            print
++            print "Failed to find the necessary bits to build these modules:"
++            print_three_column(missing)
++            print ("To find the necessary bits, look in setup.py in"
++                   " detect_modules() for the module's name.")
++            print
++
++        if self.failed:
++            failed = self.failed[:]
++            print
++            print "Failed to build these modules:"
++            print_three_column(failed)
++            print
++
++    def build_extension(self, ext):
++
++        if ext.name == '_ctypes':
++            if not self.configure_ctypes(ext):
++                return
++
++        try:
++            build_ext.build_extension(self, ext)
++        except (CCompilerError, DistutilsError), why:
++            self.announce('WARNING: building of extension "%s" failed: %s' %
++                          (ext.name, sys.exc_info()[1]))
++            self.failed.append(ext.name)
++            return
++        # Workaround for Mac OS X: The Carbon-based modules cannot be
++        # reliably imported into a command-line Python
++        if 'Carbon' in ext.extra_link_args:
++            self.announce(
++                'WARNING: skipping import check for Carbon-based "%s"' %
++                ext.name)
++            return
++
++        if self.get_platform() == 'darwin' and (
++                sys.maxint > 2**32 and '-arch' in ext.extra_link_args):
++            # Don't bother doing an import check when an extension was
++            # build with an explicit '-arch' flag on OSX. That's currently
++            # only used to build 32-bit only extensions in a 4-way
++            # universal build and loading 32-bit code into a 64-bit
++            # process will fail.
++            self.announce(
++                'WARNING: skipping import check for "%s"' %
++                ext.name)
++            return
++
++        # Workaround for Cygwin: Cygwin currently has fork issues when many
++        # modules have been imported
++        if self.get_platform() == 'cygwin':
++            self.announce('WARNING: skipping import check for Cygwin-based "%s"'
++                % ext.name)
++            return
++        ext_filename = os.path.join(
++            self.build_lib,
++            self.get_ext_filename(self.get_ext_fullname(ext.name)))
++        try:
++            imp.load_dynamic(ext.name, ext_filename)
++        except ImportError, why:
++            self.failed.append(ext.name)
++            self.announce('*** WARNING: renaming "%s" since importing it'
++                          ' failed: %s' % (ext.name, why), level=3)
++            assert not self.inplace
++            basename, tail = os.path.splitext(ext_filename)
++            newname = basename + "_failed" + tail
++            if os.path.exists(newname):
++                os.remove(newname)
++            os.rename(ext_filename, newname)
++
++            # XXX -- This relies on a Vile HACK in
++            # distutils.command.build_ext.build_extension().  The
++            # _built_objects attribute is stored there strictly for
++            # use here.
++            # If there is a failure, _built_objects may not be there,
++            # so catch the AttributeError and move on.
++            try:
++                for filename in self._built_objects:
++                    os.remove(filename)
++            except AttributeError:
++                self.announce('unable to remove files (ignored)')
++        except:
++            exc_type, why, tb = sys.exc_info()
++            self.announce('*** WARNING: importing extension "%s" '
++                          'failed with %s: %s' % (ext.name, exc_type, why),
++                          level=3)
++            self.failed.append(ext.name)
++
++    def get_platform(self):
++        # Get value of sys.platform
++        for platform in ['cygwin', 'beos', 'darwin', 'atheos', 'osf1']:
++            if sys.platform.startswith(platform):
++                return platform
++        return sys.platform
++
++    def detect_modules(self):
++        if sys.platform != 'sunos5':
++            # Ensure that /usr/local is always used
++            add_dir_to_list(self.compiler.library_dirs, '/usr/local/lib')
++            add_dir_to_list(self.compiler.include_dirs, '/usr/local/include')
++        # Add paths specified in the environment variables LDFLAGS and
++        # CPPFLAGS for header and library files.
++        # We must get the values from the Makefile and not the environment
++        # directly since an inconsistently reproducible issue comes up where
++        # the environment variable is not set even though the value were passed
++        # into configure and stored in the Makefile (issue found on OS X 10.3).
++        for env_var, arg_name, dir_list in (
++                ('LDFLAGS', '-R', self.compiler.runtime_library_dirs),
++                ('LDFLAGS', '-L', self.compiler.library_dirs),
++                ('CPPFLAGS', '-I', self.compiler.include_dirs)):
++            env_val = sysconfig.get_config_var(env_var)
++            if env_val:
++                # To prevent optparse from raising an exception about any
++                # options in env_val that it doesn't know about we strip out
++                # all double dashes and any dashes followed by a character
++                # that is not for the option we are dealing with.
++                #
++                # Please note that order of the regex is important!  We must
++                # strip out double-dashes first so that we don't end up with
++                # substituting "--Long" to "-Long" and thus lead to "ong" being
++                # used for a library directory.
++                env_val = re.sub(r'(^|\s+)-(-|(?!%s))' % arg_name[1],
++                                 ' ', env_val)
++                parser = optparse.OptionParser()
++                # Make sure that allowing args interspersed with options is
++                # allowed
++                parser.allow_interspersed_args = True
++                parser.error = lambda msg: None
++                parser.add_option(arg_name, dest="dirs", action="append")
++                options = parser.parse_args(env_val.split())[0]
++                if options.dirs:
++                    for directory in reversed(options.dirs):
++                        add_dir_to_list(dir_list, directory)
++
++        if os.path.normpath(sys.prefix) != '/usr':
++            add_dir_to_list(self.compiler.library_dirs,
++                            sysconfig.get_config_var("LIBDIR"))
++            add_dir_to_list(self.compiler.include_dirs,
++                            sysconfig.get_config_var("INCLUDEDIR"))
++
++        try:
++            have_unicode = unicode
++        except NameError:
++            have_unicode = 0
++
++        # lib_dirs and inc_dirs are used to search for files;
++        # if a file is found in one of those directories, it can
++        # be assumed that no additional -I,-L directives are needed.
++        lib_dirs = self.compiler.library_dirs + [
++            '/lib64', '/usr/lib64',
++            '/lib', '/usr/lib',
++            ]
++        inc_dirs = self.compiler.include_dirs + ['/usr/include']
++        exts = []
++        missing = []
++
++        config_h = sysconfig.get_config_h_filename()
++        config_h_vars = sysconfig.parse_config_h(open(config_h))
++
++        platform = self.get_platform()
++        (srcdir,) = sysconfig.get_config_vars('srcdir')
++
++        # Check for AtheOS which has libraries in non-standard locations
++        if platform == 'atheos':
++            lib_dirs += ['/system/libs', '/atheos/autolnk/lib']
++            lib_dirs += os.getenv('LIBRARY_PATH', '').split(os.pathsep)
++            inc_dirs += ['/system/include', '/atheos/autolnk/include']
++            inc_dirs += os.getenv('C_INCLUDE_PATH', '').split(os.pathsep)
++
++        # OSF/1 and Unixware have some stuff in /usr/ccs/lib (like -ldb)
++        if platform in ['osf1', 'unixware7', 'openunix8']:
++            lib_dirs += ['/usr/ccs/lib']
++
++        if platform == 'darwin':
++            # This should work on any unixy platform ;-)
++            # If the user has bothered specifying additional -I and -L flags
++            # in OPT and LDFLAGS we might as well use them here.
++            #   NOTE: using shlex.split would technically be more correct, but
++            # also gives a bootstrap problem. Let's hope nobody uses directories
++            # with whitespace in the name to store libraries.
++            cflags, ldflags = sysconfig.get_config_vars(
++                    'CFLAGS', 'LDFLAGS')
++            for item in cflags.split():
++                if item.startswith('-I'):
++                    inc_dirs.append(item[2:])
++
++            for item in ldflags.split():
++                if item.startswith('-L'):
++                    lib_dirs.append(item[2:])
++
++        # Check for MacOS X, which doesn't need libm.a at all
++        math_libs = ['m']
++        if platform in ['darwin', 'beos', 'mac']:
++            math_libs = []
++
++        # XXX Omitted modules: gl, pure, dl, SGI-specific modules
++
++        #
++        # The following modules are all pretty straightforward, and compile
++        # on pretty much any POSIXish platform.
++        #
++
++        # Some modules that are normally always on:
++        exts.append( Extension('_weakref', ['_weakref.c']) )
++
++        # array objects
++        exts.append( Extension('array', ['arraymodule.c']) )
++        # complex math library functions
++        exts.append( Extension('cmath', ['cmathmodule.c'],
++                               libraries=math_libs) )
++
++        # math library functions, e.g. sin()
++        exts.append( Extension('math',  ['mathmodule.c'],
++                               libraries=math_libs) )
++        # fast string operations implemented in C
++        exts.append( Extension('strop', ['stropmodule.c']) )
++        # time operations and variables
++        exts.append( Extension('time', ['timemodule.c'],
++                               libraries=math_libs) )
++        exts.append( Extension('datetime', ['datetimemodule.c', 'timemodule.c'],
++                               libraries=math_libs) )
++        # fast iterator tools implemented in C
++        exts.append( Extension("itertools", ["itertoolsmodule.c"]) )
++        # code that will be builtins in the future, but conflict with the
++        #  current builtins
++        exts.append( Extension('future_builtins', ['future_builtins.c']) )
++        # random number generator implemented in C
++        exts.append( Extension("_random", ["_randommodule.c"]) )
++        # high-performance collections
++        exts.append( Extension("_collections", ["_collectionsmodule.c"]) )
++        # bisect
++        exts.append( Extension("_bisect", ["_bisectmodule.c"]) )
++        # heapq
++        exts.append( Extension("_heapq", ["_heapqmodule.c"]) )
++        # operator.add() and similar goodies
++        exts.append( Extension('operator', ['operator.c']) )
++        # Python 3.0 _fileio module
++        exts.append( Extension("_fileio", ["_fileio.c"]) )
++        # Python 3.0 _bytesio module
++        exts.append( Extension("_bytesio", ["_bytesio.c"]) )
++        # _functools
++        exts.append( Extension("_functools", ["_functoolsmodule.c"]) )
++        # _json speedups
++        exts.append( Extension("_json", ["_json.c"]) )
++        # Python C API test module
++        exts.append( Extension('_testcapi', ['_testcapimodule.c'],
++                               depends=['testcapi_long.h']) )
++        # profilers (_lsprof is for cProfile.py)
++        exts.append( Extension('_hotshot', ['_hotshot.c']) )
++        exts.append( Extension('_lsprof', ['_lsprof.c', 'rotatingtree.c']) )
++        # static Unicode character database
++        if have_unicode:
++            exts.append( Extension('unicodedata', ['unicodedata.c']) )
++        else:
++            missing.append('unicodedata')
++        # access to ISO C locale support
++        data = open('pyconfig.h').read()
++        m = re.search(r"#s*define\s+WITH_LIBINTL\s+1\s*", data)
++        if m is not None:
++            locale_libs = ['intl']
++        else:
++            locale_libs = []
++        if platform == 'darwin':
++            locale_extra_link_args = ['-framework', 'CoreFoundation']
++        else:
++            locale_extra_link_args = []
++
++
++        exts.append( Extension('_locale', ['_localemodule.c'],
++                               libraries=locale_libs,
++                               extra_link_args=locale_extra_link_args) )
++
++        # Modules with some UNIX dependencies -- on by default:
++        # (If you have a really backward UNIX, select and socket may not be
++        # supported...)
++
++        # fcntl(2) and ioctl(2)
++        exts.append( Extension('fcntl', ['fcntlmodule.c']) )
++        if platform not in ['mac']:
++            # pwd(3)
++            exts.append( Extension('pwd', ['pwdmodule.c']) )
++            # grp(3)
++            exts.append( Extension('grp', ['grpmodule.c']) )
++            # spwd, shadow passwords
++            if (config_h_vars.get('HAVE_GETSPNAM', False) or
++                    config_h_vars.get('HAVE_GETSPENT', False)):
++                exts.append( Extension('spwd', ['spwdmodule.c']) )
++            else:
++                missing.append('spwd')
++        else:
++            missing.extend(['pwd', 'grp', 'spwd'])
++
++        # select(2); not on ancient System V
++        exts.append( Extension('select', ['selectmodule.c']) )
++
++        # Fred Drake's interface to the Python parser
++        exts.append( Extension('parser', ['parsermodule.c']) )
++
++        # cStringIO and cPickle
++        exts.append( Extension('cStringIO', ['cStringIO.c']) )
++        exts.append( Extension('cPickle', ['cPickle.c']) )
++
++        # Memory-mapped files (also works on Win32).
++        if platform not in ['atheos', 'mac']:
++            exts.append( Extension('mmap', ['mmapmodule.c']) )
++        else:
++            missing.append('mmap')
++
++        # Lance Ellinghaus's syslog module
++        if platform not in ['mac']:
++            # syslog daemon interface
++            exts.append( Extension('syslog', ['syslogmodule.c']) )
++        else:
++            missing.append('syslog')
++
++        # George Neville-Neil's timing module:
++        # Deprecated in PEP 4 http://www.python.org/peps/pep-0004.html
++        # http://mail.python.org/pipermail/python-dev/2006-January/060023.html
++        #exts.append( Extension('timing', ['timingmodule.c']) )
++
++        #
++        # Here ends the simple stuff.  From here on, modules need certain
++        # libraries, are platform-specific, or present other surprises.
++        #
++
++        # Multimedia modules
++        # These don't work for 64-bit platforms!!!
++        # These represent audio samples or images as strings:
++
++        # Operations on audio samples
++        # According to #993173, this one should actually work fine on
++        # 64-bit platforms.
++        exts.append( Extension('audioop', ['audioop.c']) )
++
++        # Disabled on 64-bit platforms
++        if sys.maxint != 9223372036854775807L:
++            # Operations on images
++            exts.append( Extension('imageop', ['imageop.c']) )
++        else:
++            missing.extend(['imageop'])
++
++        # readline
++        do_readline = self.compiler.find_library_file(lib_dirs, 'readline')
++        if platform == 'darwin': # and os.uname()[2] < '9.':
++            # MacOSX 10.4 has a broken readline. Don't try to build
++            # the readline module unless the user has installed a fixed
++            # readline package
++            # FIXME: The readline emulation on 10.5 is better, but the
++            # readline module doesn't compile out of the box.
++            if find_file('readline/rlconf.h', inc_dirs, []) is None:
++                do_readline = False
++        if do_readline:
++            if sys.platform == 'darwin':
++                # In every directory on the search path search for a dynamic
++                # library and then a static library, instead of first looking
++                # for dynamic libraries on the entiry path.
++                # This way a staticly linked custom readline gets picked up
++                # before the (broken) dynamic library in /usr/lib.
++                readline_extra_link_args = ('-Wl,-search_paths_first',)
++            else:
++                readline_extra_link_args = ()
++
++            readline_libs = ['readline']
++            if self.compiler.find_library_file(lib_dirs,
++                                                 'ncursesw'):
++                readline_libs.append('ncursesw')
++            elif self.compiler.find_library_file(lib_dirs,
++                                                 'ncurses'):
++                readline_libs.append('ncurses')
++            elif self.compiler.find_library_file(lib_dirs, 'curses'):
++                readline_libs.append('curses')
++            elif self.compiler.find_library_file(lib_dirs +
++                                               ['/usr/lib/termcap'],
++                                               'termcap'):
++                readline_libs.append('termcap')
++            exts.append( Extension('readline', ['readline.c'],
++                                   library_dirs=['/usr/lib/termcap'],
++                                   extra_link_args=readline_extra_link_args,
++                                   libraries=readline_libs) )
++        else:
++            missing.append('readline')
++
++        if platform not in ['mac']:
++            # crypt module.
++
++            if self.compiler.find_library_file(lib_dirs, 'crypt'):
++                libs = ['crypt']
++            else:
++                libs = []
++            exts.append( Extension('crypt', ['cryptmodule.c'], libraries=libs) )
++        else:
++            missing.append('crypt')
++
++        # CSV files
++        exts.append( Extension('_csv', ['_csv.c']) )
++
++        # socket(2)
++        socket_libs = []
++        if self.compiler.find_library_file(lib_dirs,
++                                           'socket'):
++            socket_libs.append('socket')
++        if self.compiler.find_library_file(lib_dirs,
++                                           'nsl'):
++            socket_libs.append('nsl')
++        if self.compiler.find_library_file(lib_dirs,
++                                           'resolv'):
++            socket_libs.append('resolv')
++        exts.append( Extension('_socket', ['socketmodule.c'],
++                               depends = ['socketmodule.h'],
++                               libraries = socket_libs) )
++        # Detect SSL support for the socket module (via _ssl)
++        search_for_ssl_incs_in = [
++                              '/usr/sfw/include',
++                              '/usr/contrib/ssl/include/'
++                             ]
++        ssl_incs = find_file('openssl/ssl.h', inc_dirs,
++                             search_for_ssl_incs_in
++                             )
++        if ssl_incs is not None:
++            krb5_h = find_file('krb5.h', inc_dirs,
++                               ['/usr/kerberos/include'])
++            if krb5_h:
++                ssl_incs += krb5_h
++        if sys.maxint == 2147483647L:
++            sfw_libdir = '/usr/sfw/lib';
++        else:
++            sfw_libdir = '/usr/sfw/lib/64';
++        ssl_libs = find_library_file(self.compiler, 'ssl',lib_dirs,
++                                     [sfw_libdir,
++                                      '/usr/contrib/ssl/lib/'
++                                     ] )
++
++        if (ssl_incs is not None and
++            ssl_libs is not None):
++            exts.append( Extension('_ssl', ['_ssl.c'],
++                                   include_dirs = ssl_incs,
++                                   library_dirs = ssl_libs,
++				   runtime_library_dirs = ssl_libs,
++                                   libraries = ['ssl', 'crypto'],
++                                   depends = ['socketmodule.h']), )
++        else:
++            missing.append('_ssl')
++
++        # find out which version of OpenSSL we have
++        openssl_ver = 0
++        openssl_ver_re = re.compile(
++            '^\s*#\s*define\s+OPENSSL_VERSION_NUMBER\s+(0x[0-9a-fA-F]+)' )
++        for ssl_inc_dir in inc_dirs + search_for_ssl_incs_in:
++            name = os.path.join(ssl_inc_dir, 'openssl', 'opensslv.h')
++            if os.path.isfile(name):
++                try:
++                    incfile = open(name, 'r')
++                    for line in incfile:
++                        m = openssl_ver_re.match(line)
++                        if m:
++                            openssl_ver = eval(m.group(1))
++                            break
++                except IOError:
++                    pass
++
++            # first version found is what we'll use (as the compiler should)
++            if openssl_ver:
++                break
++
++        #print 'openssl_ver = 0x%08x' % openssl_ver
++
++        if (ssl_incs is not None and
++            ssl_libs is not None and
++            openssl_ver >= 0x00907000):
++            # The _hashlib module wraps optimized implementations
++            # of hash functions from the OpenSSL library.
++            exts.append( Extension('_hashlib', ['_hashopenssl.c'],
++                                   include_dirs = ssl_incs,
++                                   library_dirs = ssl_libs,
++                                   libraries = ['ssl', 'crypto']) )
++            # these aren't strictly missing since they are unneeded.
++            #missing.extend(['_sha', '_md5'])
++        else:
++            # The _sha module implements the SHA1 hash algorithm.
++            exts.append( Extension('_sha', ['shamodule.c']) )
++            # The _md5 module implements the RSA Data Security, Inc. MD5
++            # Message-Digest Algorithm, described in RFC 1321.  The
++            # necessary files md5.c and md5.h are included here.
++            exts.append( Extension('_md5',
++                            sources = ['md5module.c', 'md5.c'],
++                            depends = ['md5.h']) )
++            missing.append('_hashlib')
++
++        if (openssl_ver < 0x00908000):
++            # OpenSSL doesn't do these until 0.9.8 so we'll bring our own hash
++            exts.append( Extension('_sha256', ['sha256module.c']) )
++            exts.append( Extension('_sha512', ['sha512module.c']) )
++
++        # Modules that provide persistent dictionary-like semantics.  You will
++        # probably want to arrange for at least one of them to be available on
++        # your machine, though none are defined by default because of library
++        # dependencies.  The Python module anydbm.py provides an
++        # implementation independent wrapper for these; dumbdbm.py provides
++        # similar functionality (but slower of course) implemented in Python.
++
++        # Sleepycat^WOracle Berkeley DB interface.
++        #  http://www.oracle.com/database/berkeley-db/db/index.html
++        #
++        # This requires the Sleepycat^WOracle DB code. The supported versions
++        # are set below.  Visit the URL above to download
++        # a release.  Most open source OSes come with one or more
++        # versions of BerkeleyDB already installed.
++
++        max_db_ver = (4, 7)
++        min_db_ver = (3, 3)
++        db_setup_debug = False   # verbose debug prints from this script?
++
++        def allow_db_ver(db_ver):
++            """Returns a boolean if the given BerkeleyDB version is acceptable.
++
++            Args:
++              db_ver: A tuple of the version to verify.
++            """
++            if not (min_db_ver <= db_ver <= max_db_ver):
++                return False
++            # Use this function to filter out known bad configurations.
++            if (4, 6) == db_ver[:2]:
++                # BerkeleyDB 4.6.x is not stable on many architectures.
++                arch = platform_machine()
++                if arch not in ('i386', 'i486', 'i586', 'i686',
++                                'x86_64', 'ia64'):
++                    return False
++            return True
++
++        def gen_db_minor_ver_nums(major):
++            if major == 4:
++                for x in range(max_db_ver[1]+1):
++                    if allow_db_ver((4, x)):
++                        yield x
++            elif major == 3:
++                for x in (3,):
++                    if allow_db_ver((3, x)):
++                        yield x
++            else:
++                raise ValueError("unknown major BerkeleyDB version", major)
++
++        # construct a list of paths to look for the header file in on
++        # top of the normal inc_dirs.
++        db_inc_paths = [
++            '/usr/include/db4',
++            '/usr/local/include/db4',
++            '/opt/sfw/include/db4',
++            '/usr/include/db3',
++            '/usr/local/include/db3',
++            '/opt/sfw/include/db3',
++            # Fink defaults (http://fink.sourceforge.net/)
++            '/sw/include/db4',
++            '/sw/include/db3',
++        ]
++        # 4.x minor number specific paths
++        for x in gen_db_minor_ver_nums(4):
++            db_inc_paths.append('/usr/include/db4%d' % x)
++            db_inc_paths.append('/usr/include/db4.%d' % x)
++            db_inc_paths.append('/usr/local/BerkeleyDB.4.%d/include' % x)
++            db_inc_paths.append('/usr/local/include/db4%d' % x)
++            db_inc_paths.append('/pkg/db-4.%d/include' % x)
++            db_inc_paths.append('/opt/db-4.%d/include' % x)
++            # MacPorts default (http://www.macports.org/)
++            db_inc_paths.append('/opt/local/include/db4%d' % x)
++        # 3.x minor number specific paths
++        for x in gen_db_minor_ver_nums(3):
++            db_inc_paths.append('/usr/include/db3%d' % x)
++            db_inc_paths.append('/usr/local/BerkeleyDB.3.%d/include' % x)
++            db_inc_paths.append('/usr/local/include/db3%d' % x)
++            db_inc_paths.append('/pkg/db-3.%d/include' % x)
++            db_inc_paths.append('/opt/db-3.%d/include' % x)
++
++        # Add some common subdirectories for Sleepycat DB to the list,
++        # based on the standard include directories. This way DB3/4 gets
++        # picked up when it is installed in a non-standard prefix and
++        # the user has added that prefix into inc_dirs.
++        std_variants = []
++        for dn in inc_dirs:
++            std_variants.append(os.path.join(dn, 'db3'))
++            std_variants.append(os.path.join(dn, 'db4'))
++            for x in gen_db_minor_ver_nums(4):
++                std_variants.append(os.path.join(dn, "db4%d"%x))
++                std_variants.append(os.path.join(dn, "db4.%d"%x))
++            for x in gen_db_minor_ver_nums(3):
++                std_variants.append(os.path.join(dn, "db3%d"%x))
++                std_variants.append(os.path.join(dn, "db3.%d"%x))
++
++        db_inc_paths = std_variants + db_inc_paths
++        db_inc_paths = [p for p in db_inc_paths if os.path.exists(p)]
++
++        db_ver_inc_map = {}
++
++        class db_found(Exception): pass
++        try:
++            # See whether there is a Sleepycat header in the standard
++            # search path.
++            for d in inc_dirs + db_inc_paths:
++                f = os.path.join(d, "db.h")
++                if db_setup_debug: print "db: looking for db.h in", f
++                if os.path.exists(f):
++                    f = open(f).read()
++                    m = re.search(r"#define\WDB_VERSION_MAJOR\W(\d+)", f)
++                    if m:
++                        db_major = int(m.group(1))
++                        m = re.search(r"#define\WDB_VERSION_MINOR\W(\d+)", f)
++                        db_minor = int(m.group(1))
++                        db_ver = (db_major, db_minor)
++
++                        # Avoid 4.6 prior to 4.6.21 due to a BerkeleyDB bug
++                        if db_ver == (4, 6):
++                            m = re.search(r"#define\WDB_VERSION_PATCH\W(\d+)", f)
++                            db_patch = int(m.group(1))
++                            if db_patch < 21:
++                                print "db.h:", db_ver, "patch", db_patch,
++                                print "being ignored (4.6.x must be >= 4.6.21)"
++                                continue
++
++                        if ( (not db_ver_inc_map.has_key(db_ver)) and
++                            allow_db_ver(db_ver) ):
++                            # save the include directory with the db.h version
++                            # (first occurrence only)
++                            db_ver_inc_map[db_ver] = d
++                            if db_setup_debug:
++                                print "db.h: found", db_ver, "in", d
++                        else:
++                            # we already found a header for this library version
++                            if db_setup_debug: print "db.h: ignoring", d
++                    else:
++                        # ignore this header, it didn't contain a version number
++                        if db_setup_debug:
++                            print "db.h: no version number version in", d
++
++            db_found_vers = db_ver_inc_map.keys()
++            db_found_vers.sort()
++
++            while db_found_vers:
++                db_ver = db_found_vers.pop()
++                db_incdir = db_ver_inc_map[db_ver]
++
++                # check lib directories parallel to the location of the header
++                db_dirs_to_check = [
++                    db_incdir.replace("include", 'lib64'),
++                    db_incdir.replace("include", 'lib'),
++                ]
++                db_dirs_to_check = filter(os.path.isdir, db_dirs_to_check)
++
++                # Look for a version specific db-X.Y before an ambiguoius dbX
++                # XXX should we -ever- look for a dbX name?  Do any
++                # systems really not name their library by version and
++                # symlink to more general names?
++                for dblib in (('db-%d.%d' % db_ver),
++                              ('db%d%d' % db_ver),
++                              ('db%d' % db_ver[0])):
++                    dblib_file = self.compiler.find_library_file(
++                                    db_dirs_to_check + lib_dirs, dblib )
++                    if dblib_file:
++                        dblib_dir = [ os.path.abspath(os.path.dirname(dblib_file)) ]
++                        raise db_found
++                    else:
++                        if db_setup_debug: print "db lib: ", dblib, "not found"
++
++        except db_found:
++            if db_setup_debug:
++                print "bsddb using BerkeleyDB lib:", db_ver, dblib
++                print "bsddb lib dir:", dblib_dir, " inc dir:", db_incdir
++            db_incs = [db_incdir]
++            dblibs = [dblib]
++            # We add the runtime_library_dirs argument because the
++            # BerkeleyDB lib we're linking against often isn't in the
++            # system dynamic library search path.  This is usually
++            # correct and most trouble free, but may cause problems in
++            # some unusual system configurations (e.g. the directory
++            # is on an NFS server that goes away).
++            exts.append(Extension('_bsddb', ['_bsddb.c'],
++                                  depends = ['bsddb.h'],
++                                  library_dirs=dblib_dir,
++                                  runtime_library_dirs=dblib_dir,
++                                  include_dirs=db_incs,
++                                  libraries=dblibs))
++        else:
++            if db_setup_debug: print "db: no appropriate library found"
++            db_incs = None
++            dblibs = []
++            dblib_dir = None
++            missing.append('_bsddb')
++
++        # The sqlite interface
++        sqlite_setup_debug = False   # verbose debug prints from this script?
++
++        # We hunt for #define SQLITE_VERSION "n.n.n"
++        # We need to find >= sqlite version 3.0.8
++        sqlite_incdir = sqlite_libdir = None
++        sqlite_inc_paths = [ '/usr/include',
++                             '/usr/include/sqlite',
++                             '/usr/include/sqlite3',
++                             '/usr/local/include',
++                             '/usr/local/include/sqlite',
++                             '/usr/local/include/sqlite3',
++                           ]
++        MIN_SQLITE_VERSION_NUMBER = (3, 0, 8)
++        MIN_SQLITE_VERSION = ".".join([str(x)
++                                    for x in MIN_SQLITE_VERSION_NUMBER])
++
++        # Scan the default include directories before the SQLite specific
++        # ones. This allows one to override the copy of sqlite on OSX,
++        # where /usr/include contains an old version of sqlite.
++        for d in inc_dirs + sqlite_inc_paths:
++            f = os.path.join(d, "sqlite3.h")
++            if os.path.exists(f):
++                if sqlite_setup_debug: print "sqlite: found %s"%f
++                incf = open(f).read()
++                m = re.search(
++                    r'\s*.*#\s*.*define\s.*SQLITE_VERSION\W*"(.*)"', incf)
++                if m:
++                    sqlite_version = m.group(1)
++                    sqlite_version_tuple = tuple([int(x)
++                                        for x in sqlite_version.split(".")])
++                    if sqlite_version_tuple >= MIN_SQLITE_VERSION_NUMBER:
++                        # we win!
++                        if sqlite_setup_debug:
++                            print "%s/sqlite3.h: version %s"%(d, sqlite_version)
++                        sqlite_incdir = d
++                        break
++                    else:
++                        if sqlite_setup_debug:
++                            print "%s: version %d is too old, need >= %s"%(d,
++                                        sqlite_version, MIN_SQLITE_VERSION)
++                elif sqlite_setup_debug:
++                    print "sqlite: %s had no SQLITE_VERSION"%(f,)
++
++        if sqlite_incdir:
++            sqlite_dirs_to_check = [
++                os.path.join(sqlite_incdir, '..', 'lib64'),
++                os.path.join(sqlite_incdir, '..', 'lib'),
++                os.path.join(sqlite_incdir, '..', '..', 'lib64'),
++                os.path.join(sqlite_incdir, '..', '..', 'lib'),
++            ]
++            sqlite_libfile = self.compiler.find_library_file(
++                                sqlite_dirs_to_check + lib_dirs, 'sqlite3')
++            if sqlite_libfile:
++                sqlite_libdir = [os.path.abspath(os.path.dirname(sqlite_libfile))]
++
++        if sqlite_incdir and sqlite_libdir:
++            sqlite_srcs = ['_sqlite/cache.c',
++                '_sqlite/connection.c',
++                '_sqlite/cursor.c',
++                '_sqlite/microprotocols.c',
++                '_sqlite/module.c',
++                '_sqlite/prepare_protocol.c',
++                '_sqlite/row.c',
++                '_sqlite/statement.c',
++                '_sqlite/util.c', ]
++
++            sqlite_defines = []
++            if sys.platform != "win32":
++                sqlite_defines.append(('MODULE_NAME', '"sqlite3"'))
++            else:
++                sqlite_defines.append(('MODULE_NAME', '\\"sqlite3\\"'))
++
++
++            if sys.platform == 'darwin':
++                # In every directory on the search path search for a dynamic
++                # library and then a static library, instead of first looking
++                # for dynamic libraries on the entiry path.
++                # This way a staticly linked custom sqlite gets picked up
++                # before the dynamic library in /usr/lib.
++                sqlite_extra_link_args = ('-Wl,-search_paths_first',)
++            else:
++                sqlite_extra_link_args = ()
++
++            exts.append(Extension('_sqlite3', sqlite_srcs,
++                                  define_macros=sqlite_defines,
++                                  include_dirs=["Modules/_sqlite",
++                                                sqlite_incdir],
++                                  library_dirs=sqlite_libdir,
++                                  runtime_library_dirs=sqlite_libdir,
++                                  extra_link_args=sqlite_extra_link_args,
++                                  libraries=["sqlite3",]))
++        else:
++            missing.append('_sqlite3')
++
++        # Look for Berkeley db 1.85.   Note that it is built as a different
++        # module name so it can be included even when later versions are
++        # available.  A very restrictive search is performed to avoid
++        # accidentally building this module with a later version of the
++        # underlying db library.  May BSD-ish Unixes incorporate db 1.85
++        # symbols into libc and place the include file in /usr/include.
++        #
++        # If the better bsddb library can be built (db_incs is defined)
++        # we do not build this one.  Otherwise this build will pick up
++        # the more recent berkeleydb's db.h file first in the include path
++        # when attempting to compile and it will fail.
++        f = "/usr/include/db.h"
++        if os.path.exists(f) and not db_incs:
++            data = open(f).read()
++            m = re.search(r"#s*define\s+HASHVERSION\s+2\s*", data)
++            if m is not None:
++                # bingo - old version used hash file format version 2
++                ### XXX this should be fixed to not be platform-dependent
++                ### but I don't have direct access to an osf1 platform and
++                ### seemed to be muffing the search somehow
++                libraries = platform == "osf1" and ['db'] or None
++                if libraries is not None:
++                    exts.append(Extension('bsddb185', ['bsddbmodule.c'],
++                                          libraries=libraries))
++                else:
++                    exts.append(Extension('bsddb185', ['bsddbmodule.c']))
++            else:
++                missing.append('bsddb185')
++        else:
++            missing.append('bsddb185')
++
++        # The standard Unix dbm module:
++        if platform not in ['cygwin']:
++            if find_file("ndbm.h", inc_dirs, []) is not None:
++                # Some systems have -lndbm, others don't
++                if self.compiler.find_library_file(lib_dirs, 'ndbm'):
++                    ndbm_libs = ['ndbm']
++                else:
++                    ndbm_libs = []
++                exts.append( Extension('dbm', ['dbmmodule.c'],
++                                       define_macros=[('HAVE_NDBM_H',None)],
++                                       libraries = ndbm_libs ) )
++            elif self.compiler.find_library_file(lib_dirs, 'gdbm'):
++                gdbm_libs = ['gdbm']
++                if self.compiler.find_library_file(lib_dirs, 'gdbm_compat'):
++                    gdbm_libs.append('gdbm_compat')
++                if find_file("gdbm/ndbm.h", inc_dirs, []) is not None:
++                    exts.append( Extension(
++                        'dbm', ['dbmmodule.c'],
++                        define_macros=[('HAVE_GDBM_NDBM_H',None)],
++                        libraries = gdbm_libs ) )
++                elif find_file("gdbm-ndbm.h", inc_dirs, []) is not None:
++                    exts.append( Extension(
++                        'dbm', ['dbmmodule.c'],
++                        define_macros=[('HAVE_GDBM_DASH_NDBM_H',None)],
++                        libraries = gdbm_libs ) )
++                else:
++                    missing.append('dbm')
++            elif db_incs is not None:
++                exts.append( Extension('dbm', ['dbmmodule.c'],
++                                       library_dirs=dblib_dir,
++                                       runtime_library_dirs=dblib_dir,
++                                       include_dirs=db_incs,
++                                       define_macros=[('HAVE_BERKDB_H',None),
++                                                      ('DB_DBM_HSEARCH',None)],
++                                       libraries=dblibs))
++            else:
++                missing.append('dbm')
++
++        # Anthony Baxter's gdbm module.  GNU dbm(3) will require -lgdbm:
++        if (self.compiler.find_library_file(lib_dirs, 'gdbm')):
++            exts.append( Extension('gdbm', ['gdbmmodule.c'],
++                                   libraries = ['gdbm'] ) )
++        else:
++            missing.append('gdbm')
++
++        # Unix-only modules
++        if platform not in ['mac', 'win32']:
++            # Steen Lumholt's termios module
++            exts.append( Extension('termios', ['termios.c']) )
++            # Jeremy Hylton's rlimit interface
++            if platform not in ['atheos']:
++                exts.append( Extension('resource', ['resource.c']) )
++            else:
++                missing.append('resource')
++
++            # Sun yellow pages. Some systems have the functions in libc.
++            if platform not in ['cygwin', 'atheos', 'qnx6']:
++                if (self.compiler.find_library_file(lib_dirs, 'nsl')):
++                    libs = ['nsl']
++                else:
++                    libs = []
++                exts.append( Extension('nis', ['nismodule.c'],
++                                       libraries = libs) )
++            else:
++                missing.append('nis')
++        else:
++            missing.extend(['nis', 'resource', 'termios'])
++
++        # Curses support, requiring the System V version of curses, often
++        # provided by the ncurses library.
++        curses_lib_dirs = []
++        curses_inc_dirs = []
++        if platform == 'sunos5':
++            # look for ncurses in /usr/gnu on Solaris
++            curses_inc_dirs.append('/usr/include/ncurses')
++            curses_lib_dirs.append('/usr/gnu/lib')
++            curses_lib_dirs.append('/usr/gnu/lib/amd64')
++            curses_lib_dirs.append('/usr/gnu/lib/sparcv9')
++        panel_library = 'panel'
++        if (self.compiler.find_library_file(lib_dirs, 'ncursesw')):
++            curses_libs = ['ncursesw']
++            # Bug 1464056: If _curses.so links with ncursesw,
++            # _curses_panel.so must link with panelw.
++            panel_library = 'panelw'
++            exts.append( Extension('_curses', ['_cursesmodule.c'],
++                                   libraries = curses_libs) )
++        elif (self.compiler.find_library_file(lib_dirs + curses_lib_dirs, 'ncurses')):
++            curses_libs = ['ncurses']
++            exts.append( Extension('_curses', ['_cursesmodule.c'],
++                                   libraries = curses_libs,
++                                   library_dirs = curses_lib_dirs,
++                                   runtime_library_dirs = curses_lib_dirs,
++                                   include_dirs = curses_inc_dirs ) )
++        elif (self.compiler.find_library_file(lib_dirs, 'curses')
++              and platform != 'darwin'):
++                # OSX has an old Berkeley curses, not good enough for
++                # the _curses module.
++            if (self.compiler.find_library_file(lib_dirs, 'terminfo')):
++                curses_libs = ['curses', 'terminfo']
++            elif (self.compiler.find_library_file(lib_dirs, 'termcap')):
++                curses_libs = ['curses', 'termcap']
++            else:
++                curses_libs = ['curses']
++
++            exts.append( Extension('_curses', ['_cursesmodule.c'],
++                                   libraries = curses_libs) )
++        else:
++            missing.append('_curses')
++
++        # If the curses module is enabled, check for the panel module
++        if (module_enabled(exts, '_curses') and
++            self.compiler.find_library_file(lib_dirs + curses_lib_dirs, panel_library)):
++            exts.append( Extension('_curses_panel', ['_curses_panel.c'],
++                                   libraries = [panel_library] + curses_libs,
++                                   include_dirs = curses_inc_dirs,
++                                   library_dirs = curses_lib_dirs,
++                                   runtime_library_dirs = curses_lib_dirs ) )
++        else:
++            missing.append('_curses_panel')
++
++        # Andrew Kuchling's zlib module.  Note that some versions of zlib
++        # 1.1.3 have security problems.  See CERT Advisory CA-2002-07:
++        # http://www.cert.org/advisories/CA-2002-07.html
++        #
++        # zlib 1.1.4 is fixed, but at least one vendor (RedHat) has decided to
++        # patch its zlib 1.1.3 package instead of upgrading to 1.1.4.  For
++        # now, we still accept 1.1.3, because we think it's difficult to
++        # exploit this in Python, and we'd rather make it RedHat's problem
++        # than our problem <wink>.
++        #
++        # You can upgrade zlib to version 1.1.4 yourself by going to
++        # http://www.gzip.org/zlib/
++        zlib_inc = find_file('zlib.h', [], inc_dirs)
++        have_zlib = False
++        if zlib_inc is not None:
++            zlib_h = zlib_inc[0] + '/zlib.h'
++            version = '"0.0.0"'
++            version_req = '"1.1.3"'
++            fp = open(zlib_h)
++            while 1:
++                line = fp.readline()
++                if not line:
++                    break
++                if line.startswith('#define ZLIB_VERSION'):
++                    version = line.split()[2]
++                    break
++            if version >= version_req:
++                if (self.compiler.find_library_file(lib_dirs, 'z')):
++                    if sys.platform == "darwin":
++                        zlib_extra_link_args = ('-Wl,-search_paths_first',)
++                    else:
++                        zlib_extra_link_args = ()
++                    exts.append( Extension('zlib', ['zlibmodule.c'],
++                                           libraries = ['z'],
++                                           extra_link_args = zlib_extra_link_args))
++                    have_zlib = True
++                else:
++                    missing.append('zlib')
++            else:
++                missing.append('zlib')
++        else:
++            missing.append('zlib')
++
++        # Helper module for various ascii-encoders.  Uses zlib for an optimized
++        # crc32 if we have it.  Otherwise binascii uses its own.
++        if have_zlib:
++            extra_compile_args = ['-DUSE_ZLIB_CRC32']
++            libraries = ['z']
++            extra_link_args = zlib_extra_link_args
++        else:
++            extra_compile_args = []
++            libraries = []
++            extra_link_args = []
++        exts.append( Extension('binascii', ['binascii.c'],
++                               extra_compile_args = extra_compile_args,
++                               libraries = libraries,
++                               extra_link_args = extra_link_args) )
++
++        # Gustavo Niemeyer's bz2 module.
++        if (self.compiler.find_library_file(lib_dirs, 'bz2')):
++            if sys.platform == "darwin":
++                bz2_extra_link_args = ('-Wl,-search_paths_first',)
++            else:
++                bz2_extra_link_args = ()
++            exts.append( Extension('bz2', ['bz2module.c'],
++                                   libraries = ['bz2'],
++                                   extra_link_args = bz2_extra_link_args) )
++        else:
++            missing.append('bz2')
++
++        # Interface to the Expat XML parser
++        #
++        # Expat was written by James Clark and is now maintained by a
++        # group of developers on SourceForge; see www.libexpat.org for
++        # more information.  The pyexpat module was written by Paul
++        # Prescod after a prototype by Jack Jansen.  The Expat source
++        # is included in Modules/expat/.  Usage of a system
++        # shared libexpat.so/expat.dll is not advised.
++        #
++        # More information on Expat can be found at www.libexpat.org.
++        #
++        expatinc = os.path.join(os.getcwd(), srcdir, 'Modules', 'expat')
++        define_macros = [
++            ('HAVE_EXPAT_CONFIG_H', '1'),
++        ]
++
++        exts.append(Extension('pyexpat',
++                              define_macros = define_macros,
++                              include_dirs = [expatinc],
++                              sources = ['pyexpat.c',
++                                         'expat/xmlparse.c',
++                                         'expat/xmlrole.c',
++                                         'expat/xmltok.c',
++                                         ],
++                              ))
++
++        # Fredrik Lundh's cElementTree module.  Note that this also
++        # uses expat (via the CAPI hook in pyexpat).
++
++        if os.path.isfile(os.path.join(srcdir, 'Modules', '_elementtree.c')):
++            define_macros.append(('USE_PYEXPAT_CAPI', None))
++            exts.append(Extension('_elementtree',
++                                  define_macros = define_macros,
++                                  include_dirs = [expatinc],
++                                  sources = ['_elementtree.c'],
++                                  ))
++        else:
++            missing.append('_elementtree')
++
++        # Hye-Shik Chang's CJKCodecs modules.
++        if have_unicode:
++            exts.append(Extension('_multibytecodec',
++                                  ['cjkcodecs/multibytecodec.c']))
++            for loc in ('kr', 'jp', 'cn', 'tw', 'hk', 'iso2022'):
++                exts.append(Extension('_codecs_%s' % loc,
++                                      ['cjkcodecs/_codecs_%s.c' % loc]))
++        else:
++            missing.append('_multibytecodec')
++            for loc in ('kr', 'jp', 'cn', 'tw', 'hk', 'iso2022'):
++                missing.append('_codecs_%s' % loc)
++
++        # Dynamic loading module
++        if sys.maxint == 0x7fffffff:
++            # This requires sizeof(int) == sizeof(long) == sizeof(char*)
++            dl_inc = find_file('dlfcn.h', [], inc_dirs)
++            if (dl_inc is not None) and (platform not in ['atheos']):
++                exts.append( Extension('dl', ['dlmodule.c']) )
++            else:
++                missing.append('dl')
++        else:
++            missing.append('dl')
++
++        # Thomas Heller's _ctypes module
++        self.detect_ctypes(inc_dirs, lib_dirs)
++
++        # Richard Oudkerk's multiprocessing module
++        if platform == 'win32':             # Windows
++            macros = dict()
++            libraries = ['ws2_32']
++
++        elif platform == 'darwin':          # Mac OSX
++            macros = dict(
++                HAVE_SEM_OPEN=1,
++                HAVE_SEM_TIMEDWAIT=0,
++                HAVE_FD_TRANSFER=1,
++                HAVE_BROKEN_SEM_GETVALUE=1
++                )
++            libraries = []
++
++        elif platform == 'cygwin':          # Cygwin
++            macros = dict(
++                HAVE_SEM_OPEN=1,
++                HAVE_SEM_TIMEDWAIT=1,
++                HAVE_FD_TRANSFER=0,
++                HAVE_BROKEN_SEM_UNLINK=1
++                )
++            libraries = []
++
++        elif platform in ('freebsd4', 'freebsd5', 'freebsd6', 'freebsd7', 'freebsd8'):
++            # FreeBSD's P1003.1b semaphore support is very experimental
++            # and has many known problems. (as of June 2008)
++            macros = dict(                  # FreeBSD
++                HAVE_SEM_OPEN=0,
++                HAVE_SEM_TIMEDWAIT=0,
++                HAVE_FD_TRANSFER=1,
++                )
++            libraries = []
++
++        elif platform.startswith('openbsd'):
++            macros = dict(                  # OpenBSD
++                HAVE_SEM_OPEN=0,            # Not implemented
++                HAVE_SEM_TIMEDWAIT=0,
++                HAVE_FD_TRANSFER=1,
++                )
++            libraries = []
++
++        elif platform.startswith('netbsd'):
++            macros = dict(                  # at least NetBSD 5
++                HAVE_SEM_OPEN=1,
++                HAVE_SEM_TIMEDWAIT=0,
++                HAVE_FD_TRANSFER=1,
++                HAVE_BROKEN_SEM_GETVALUE=1
++                )
++            libraries = []
++
++        else:                                   # Linux and other unices
++            macros = dict(
++                HAVE_SEM_OPEN=1,
++                HAVE_SEM_TIMEDWAIT=1,
++                HAVE_FD_TRANSFER=1
++                )
++            libraries = ['rt']
++
++        if platform == 'win32':
++            multiprocessing_srcs = [ '_multiprocessing/multiprocessing.c',
++                                     '_multiprocessing/semaphore.c',
++                                     '_multiprocessing/pipe_connection.c',
++                                     '_multiprocessing/socket_connection.c',
++                                     '_multiprocessing/win32_functions.c'
++                                   ]
++
++        else:
++            multiprocessing_srcs = [ '_multiprocessing/multiprocessing.c',
++                                     '_multiprocessing/socket_connection.c'
++                                   ]
++
++            if macros.get('HAVE_SEM_OPEN', False):
++                multiprocessing_srcs.append('_multiprocessing/semaphore.c')
++
++        multiproc_libs = []
++        if platform == 'sunos5':
++            multiproc_libs = [ "xnet" ]
++
++        exts.append ( Extension('_multiprocessing', multiprocessing_srcs,
++                                 define_macros=macros.items(),
++                                 libraries=multiproc_libs,
++                                 include_dirs=["Modules/_multiprocessing"]))
++        # End multiprocessing
++
++
++        # Platform-specific libraries
++        if platform == 'linux2':
++            # Linux-specific modules
++            exts.append( Extension('linuxaudiodev', ['linuxaudiodev.c']) )
++        else:
++            missing.append('linuxaudiodev')
++
++        if platform in ('linux2', 'freebsd4', 'freebsd5', 'freebsd6',
++                        'freebsd7', 'freebsd8'):
++            exts.append( Extension('ossaudiodev', ['ossaudiodev.c']) )
++        else:
++            missing.append('ossaudiodev')
++
++        if platform == 'sunos5':
++            # SunOS specific modules
++            exts.append( Extension('sunaudiodev', ['sunaudiodev.c']) )
++        else:
++            missing.append('sunaudiodev')
++
++        if platform == 'darwin':
++            # _scproxy
++            exts.append(Extension("_scproxy", [os.path.join(srcdir, "Mac/Modules/_scproxy.c")],
++                extra_link_args= [
++                    '-framework', 'SystemConfiguration',
++                    '-framework', 'CoreFoundation'
++                ]))
++
++
++        if platform == 'darwin' and ("--disable-toolbox-glue" not in
++                sysconfig.get_config_var("CONFIG_ARGS")):
++
++            if os.uname()[2] > '8.':
++                # We're on Mac OS X 10.4 or later, the compiler should
++                # support '-Wno-deprecated-declarations'. This will
++                # surpress deprecation warnings for the Carbon extensions,
++                # these extensions wrap the Carbon APIs and even those
++                # parts that are deprecated.
++                carbon_extra_compile_args = ['-Wno-deprecated-declarations']
++            else:
++                carbon_extra_compile_args = []
++
++            # Mac OS X specific modules.
++            def macSrcExists(name1, name2=''):
++                if not name1:
++                    return None
++                names = (name1,)
++                if name2:
++                    names = (name1, name2)
++                path = os.path.join(srcdir, 'Mac', 'Modules', *names)
++                return os.path.exists(path)
++
++            def addMacExtension(name, kwds, extra_srcs=[]):
++                dirname = ''
++                if name[0] == '_':
++                    dirname = name[1:].lower()
++                cname = name + '.c'
++                cmodulename = name + 'module.c'
++                # Check for NNN.c, NNNmodule.c, _nnn/NNN.c, _nnn/NNNmodule.c
++                if macSrcExists(cname):
++                    srcs = [cname]
++                elif macSrcExists(cmodulename):
++                    srcs = [cmodulename]
++                elif macSrcExists(dirname, cname):
++                    # XXX(nnorwitz): If all the names ended with module, we
++                    # wouldn't need this condition.  ibcarbon is the only one.
++                    srcs = [os.path.join(dirname, cname)]
++                elif macSrcExists(dirname, cmodulename):
++                    srcs = [os.path.join(dirname, cmodulename)]
++                else:
++                    raise RuntimeError("%s not found" % name)
++
++                # Here's the whole point:  add the extension with sources
++                exts.append(Extension(name, srcs + extra_srcs, **kwds))
++
++            # Core Foundation
++            core_kwds = {'extra_compile_args': carbon_extra_compile_args,
++                         'extra_link_args': ['-framework', 'CoreFoundation'],
++                        }
++            addMacExtension('_CF', core_kwds, ['cf/pycfbridge.c'])
++            addMacExtension('autoGIL', core_kwds)
++
++
++
++            # Carbon
++            carbon_kwds = {'extra_compile_args': carbon_extra_compile_args,
++                           'extra_link_args': ['-framework', 'Carbon'],
++                          }
++            CARBON_EXTS = ['ColorPicker', 'gestalt', 'MacOS', 'Nav',
++                           'OSATerminology', 'icglue',
++                           # All these are in subdirs
++                           '_AE', '_AH', '_App', '_CarbonEvt', '_Cm', '_Ctl',
++                           '_Dlg', '_Drag', '_Evt', '_File', '_Folder', '_Fm',
++                           '_Help', '_Icn', '_IBCarbon', '_List',
++                           '_Menu', '_Mlte', '_OSA', '_Res', '_Qd', '_Qdoffs',
++                           '_Scrap', '_Snd', '_TE',
++                          ]
++            for name in CARBON_EXTS:
++                addMacExtension(name, carbon_kwds)
++
++            # Workaround for a bug in the version of gcc shipped with Xcode 3.
++            # The _Win extension should build just like the other Carbon extensions, but
++            # this actually results in a hard crash of the linker.
++            #
++            if '-arch ppc64' in cflags and '-arch ppc' in cflags:
++                win_kwds = {'extra_compile_args': carbon_extra_compile_args + ['-arch', 'i386', '-arch', 'ppc'],
++                               'extra_link_args': ['-framework', 'Carbon', '-arch', 'i386', '-arch', 'ppc'],
++                           }
++                addMacExtension('_Win', win_kwds)
++            else:
++                addMacExtension('_Win', carbon_kwds)
++
++
++            # Application Services & QuickTime
++            app_kwds = {'extra_compile_args': carbon_extra_compile_args,
++                        'extra_link_args': ['-framework','ApplicationServices'],
++                       }
++            addMacExtension('_Launch', app_kwds)
++            addMacExtension('_CG', app_kwds)
++
++            exts.append( Extension('_Qt', ['qt/_Qtmodule.c'],
++                        extra_compile_args=carbon_extra_compile_args,
++                        extra_link_args=['-framework', 'QuickTime',
++                                     '-framework', 'Carbon']) )
++
++
++        self.extensions.extend(exts)
++
++        # Call the method for detecting whether _tkinter can be compiled
++        self.detect_tkinter(inc_dirs, lib_dirs)
++
++        if '_tkinter' not in [e.name for e in self.extensions]:
++            missing.append('_tkinter')
++
++        return missing
++
++    def detect_tkinter_darwin(self, inc_dirs, lib_dirs):
++        # The _tkinter module, using frameworks. Since frameworks are quite
++        # different the UNIX search logic is not sharable.
++        from os.path import join, exists
++        framework_dirs = [
++            '/Library/Frameworks',
++            '/System/Library/Frameworks/',
++            join(os.getenv('HOME'), '/Library/Frameworks')
++        ]
++
++        # Find the directory that contains the Tcl.framework and Tk.framework
++        # bundles.
++        # XXX distutils should support -F!
++        for F in framework_dirs:
++            # both Tcl.framework and Tk.framework should be present
++            for fw in 'Tcl', 'Tk':
++                if not exists(join(F, fw + '.framework')):
++                    break
++            else:
++                # ok, F is now directory with both frameworks. Continure
++                # building
++                break
++        else:
++            # Tk and Tcl frameworks not found. Normal "unix" tkinter search
++            # will now resume.
++            return 0
++
++        # For 8.4a2, we must add -I options that point inside the Tcl and Tk
++        # frameworks. In later release we should hopefully be able to pass
++        # the -F option to gcc, which specifies a framework lookup path.
++        #
++        include_dirs = [
++            join(F, fw + '.framework', H)
++            for fw in 'Tcl', 'Tk'
++            for H in 'Headers', 'Versions/Current/PrivateHeaders'
++        ]
++
++        # For 8.4a2, the X11 headers are not included. Rather than include a
++        # complicated search, this is a hard-coded path. It could bail out
++        # if X11 libs are not found...
++        include_dirs.append('/usr/X11R6/include')
++        frameworks = ['-framework', 'Tcl', '-framework', 'Tk']
++
++        # All existing framework builds of Tcl/Tk don't support 64-bit
++        # architectures.
++        cflags = sysconfig.get_config_vars('CFLAGS')[0]
++        archs = re.findall('-arch\s+(\w+)', cflags)
++        fp = os.popen("file %s/Tk.framework/Tk | grep 'for architecture'"%(F,))
++        detected_archs = []
++        for ln in fp:
++            a = ln.split()[-1]
++            if a in archs:
++                detected_archs.append(ln.split()[-1])
++        fp.close()
++
++        for a in detected_archs:
++            frameworks.append('-arch')
++            frameworks.append(a)
++
++        ext = Extension('_tkinter', ['_tkinter.c', 'tkappinit.c'],
++                        define_macros=[('WITH_APPINIT', 1)],
++                        include_dirs = include_dirs,
++                        libraries = [],
++                        extra_compile_args = frameworks[2:],
++                        extra_link_args = frameworks,
++                        )
++        self.extensions.append(ext)
++        return 1
++
++
++    def detect_tkinter(self, inc_dirs, lib_dirs):
++        # The _tkinter module.
++
++        # Rather than complicate the code below, detecting and building
++        # AquaTk is a separate method. Only one Tkinter will be built on
++        # Darwin - either AquaTk, if it is found, or X11 based Tk.
++        platform = self.get_platform()
++        if (platform == 'darwin' and
++            self.detect_tkinter_darwin(inc_dirs, lib_dirs)):
++            return
++
++        # Assume we haven't found any of the libraries or include files
++        # The versions with dots are used on Unix, and the versions without
++        # dots on Windows, for detection by cygwin.
++        added_lib_dirs = []
++        tcl_tk_lib_dirs = ['/usr/sfw/lib']
++        tcl_tk_inc_dirs = ['/usr/sfw/include']
++        tcllib = tklib = tcl_includes = tk_includes = None
++        for version in ['8.5', '85', '8.4', '84', '8.3', '83', '8.2',
++                        '82', '8.1', '81', '8.0', '80']:
++            tklib = self.compiler.find_library_file(lib_dirs, 'tk' + version, tcl_tk_lib_dirs)
++            tcllib = self.compiler.find_library_file(lib_dirs, 'tcl' + version, tcl_tk_lib_dirs)
++            if tklib and tcllib:
++                # Exit the loop when we've found the Tcl/Tk libraries
++                break
++
++            tklib = self.compiler.find_library_file(tcl_tk_lib_dirs, 'tk' + version)
++            tcllib = self.compiler.find_library_file(tcl_tk_lib_dirs, 'tcl' + version)
++            if tklib and tcllib:
++                # found the libs in a non-standard dir
++                added_lib_dirs.append(os.path.dirname(tcllib))
++                # Exit the loop when we've found the Tcl/Tk libraries
++                break
++
++        # Now check for the header files
++        if tklib and tcllib:
++            # Check for the include files on Debian and {Free,Open}BSD, where
++            # they're put in /usr/include/{tcl,tk}X.Y
++            dotversion = version
++            if '.' not in dotversion and "bsd" in sys.platform.lower():
++                # OpenBSD and FreeBSD use Tcl/Tk library names like libtcl83.a,
++                # but the include subdirs are named like .../include/tcl8.3.
++                dotversion = dotversion[:-1] + '.' + dotversion[-1]
++            tcl_include_sub = []
++            tk_include_sub = []
++            for dir in inc_dirs:
++                tcl_include_sub += [dir + os.sep + "tcl" + dotversion]
++                tk_include_sub += [dir + os.sep + "tk" + dotversion]
++            tcl_include_sub += tcl_tk_inc_dirs
++            tk_include_sub += tcl_include_sub
++            tcl_includes = find_file('tcl.h', inc_dirs, tcl_include_sub)
++            tk_includes = find_file('tk.h', inc_dirs, tk_include_sub)
++
++        if (tcllib is None or tklib is None or
++            tcl_includes is None or tk_includes is None):
++            self.announce("INFO: Can't locate Tcl/Tk libs and/or headers", 2)
++            return
++
++        # OK... everything seems to be present for Tcl/Tk.
++
++        include_dirs = [] ; libs = [] ; defs = [] ; added_lib_dirs = []
++        for dir in tcl_includes + tk_includes:
++            if dir not in include_dirs:
++                include_dirs.append(dir)
++
++        # Check for various platform-specific directories
++        if platform == 'sunos5':
++            include_dirs.append('/usr/openwin/include')
++            added_lib_dirs.append('/usr/openwin/lib')
++        elif os.path.exists('/usr/X11R6/include'):
++            include_dirs.append('/usr/X11R6/include')
++            added_lib_dirs.append('/usr/X11R6/lib64')
++            added_lib_dirs.append('/usr/X11R6/lib')
++        elif os.path.exists('/usr/X11R5/include'):
++            include_dirs.append('/usr/X11R5/include')
++            added_lib_dirs.append('/usr/X11R5/lib')
++        else:
++            # Assume default location for X11
++            include_dirs.append('/usr/X11/include')
++            added_lib_dirs.append('/usr/X11/lib')
++
++        # If Cygwin, then verify that X is installed before proceeding
++        if platform == 'cygwin':
++            x11_inc = find_file('X11/Xlib.h', [], include_dirs)
++            if x11_inc is None:
++                return
++
++        # Check for BLT extension
++        if self.compiler.find_library_file(lib_dirs + added_lib_dirs,
++                                           'BLT8.0'):
++            defs.append( ('WITH_BLT', 1) )
++            libs.append('BLT8.0')
++        elif self.compiler.find_library_file(lib_dirs + added_lib_dirs,
++                                           'BLT'):
++            defs.append( ('WITH_BLT', 1) )
++            libs.append('BLT')
++
++        # Add the Tcl/Tk libraries
++        libs.append('tk'+ version)
++        libs.append('tcl'+ version)
++
++        if platform in ['aix3', 'aix4']:
++            libs.append('ld')
++
++        # Finally, link with the X11 libraries (not appropriate on cygwin)
++        if platform != "cygwin":
++            libs.append('X11')
++
++        ext = Extension('_tkinter', ['_tkinter.c', 'tkappinit.c'],
++                        define_macros=[('WITH_APPINIT', 1)] + defs,
++                        include_dirs = include_dirs,
++                        libraries = libs,
++                        library_dirs = added_lib_dirs,
++                        runtime_library_dirs = added_lib_dirs
++                        )
++        self.extensions.append(ext)
++
++##         # Uncomment these lines if you want to play with xxmodule.c
++##         ext = Extension('xx', ['xxmodule.c'])
++##         self.extensions.append(ext)
++
++        # XXX handle these, but how to detect?
++        # *** Uncomment and edit for PIL (TkImaging) extension only:
++        #       -DWITH_PIL -I../Extensions/Imaging/libImaging  tkImaging.c \
++        # *** Uncomment and edit for TOGL extension only:
++        #       -DWITH_TOGL togl.c \
++        # *** Uncomment these for TOGL extension only:
++        #       -lGL -lGLU -lXext -lXmu \
++
++    def configure_ctypes_darwin(self, ext):
++        # Darwin (OS X) uses preconfigured files, in
++        # the Modules/_ctypes/libffi_osx directory.
++        (srcdir,) = sysconfig.get_config_vars('srcdir')
++        ffi_srcdir = os.path.abspath(os.path.join(srcdir, 'Modules',
++                                                  '_ctypes', 'libffi_osx'))
++        sources = [os.path.join(ffi_srcdir, p)
++                   for p in ['ffi.c',
++                             'x86/darwin64.S',
++                             'x86/x86-darwin.S',
++                             'x86/x86-ffi_darwin.c',
++                             'x86/x86-ffi64.c',
++                             'powerpc/ppc-darwin.S',
++                             'powerpc/ppc-darwin_closure.S',
++                             'powerpc/ppc-ffi_darwin.c',
++                             'powerpc/ppc64-darwin_closure.S',
++                             ]]
++
++        # Add .S (preprocessed assembly) to C compiler source extensions.
++        self.compiler.src_extensions.append('.S')
++
++        include_dirs = [os.path.join(ffi_srcdir, 'include'),
++                        os.path.join(ffi_srcdir, 'powerpc')]
++        ext.include_dirs.extend(include_dirs)
++        ext.sources.extend(sources)
++        return True
++
++    def configure_ctypes(self, ext):
++        if not self.use_system_libffi:
++            if sys.platform == 'darwin':
++                return self.configure_ctypes_darwin(ext)
++
++            (srcdir,) = sysconfig.get_config_vars('srcdir')
++            ffi_builddir = os.path.join(self.build_temp, 'libffi')
++            ffi_srcdir = os.path.abspath(os.path.join(srcdir, 'Modules',
++                                         '_ctypes', 'libffi'))
++            ffi_configfile = os.path.join(ffi_builddir, 'fficonfig.py')
++
++            from distutils.dep_util import newer_group
++
++            config_sources = [os.path.join(ffi_srcdir, fname)
++                              for fname in os.listdir(ffi_srcdir)
++                              if os.path.isfile(os.path.join(ffi_srcdir, fname))]
++            if self.force or newer_group(config_sources,
++                                         ffi_configfile):
++                from distutils.dir_util import mkpath
++                mkpath(ffi_builddir)
++                config_args = []
++
++                # Pass empty CFLAGS because we'll just append the resulting
++                # CFLAGS to Python's; -g or -O2 is to be avoided.
++                cmd = "cd %s && env CFLAGS='' '%s/configure' %s" \
++                      % (ffi_builddir, ffi_srcdir, " ".join(config_args))
++
++                res = os.system(cmd)
++                if res or not os.path.exists(ffi_configfile):
++                    print "Failed to configure _ctypes module"
++                    return False
++
++            fficonfig = {}
++            execfile(ffi_configfile, globals(), fficonfig)
++            ffi_srcdir = os.path.join(fficonfig['ffi_srcdir'], 'src')
++
++            # Add .S (preprocessed assembly) to C compiler source extensions.
++            self.compiler.src_extensions.append('.S')
++
++            include_dirs = [os.path.join(ffi_builddir, 'include'),
++                            ffi_builddir, ffi_srcdir]
++            extra_compile_args = fficonfig['ffi_cflags'].split()
++
++            ext.sources.extend(fficonfig['ffi_sources'])
++            ext.include_dirs.extend(include_dirs)
++            ext.extra_compile_args.extend(extra_compile_args)
++        return True
++
++    def detect_ctypes(self, inc_dirs, lib_dirs):
++        self.use_system_libffi = False
++        include_dirs = []
++        extra_compile_args = []
++        extra_link_args = []
++        sources = ['_ctypes/_ctypes.c',
++                   '_ctypes/callbacks.c',
++                   '_ctypes/callproc.c',
++                   '_ctypes/stgdict.c',
++                   '_ctypes/cfield.c',
++                   '_ctypes/malloc_closure.c']
++        depends = ['_ctypes/ctypes.h']
++
++        if sys.platform == 'darwin':
++            sources.append('_ctypes/darwin/dlfcn_simple.c')
++            extra_compile_args.append('-DMACOSX')
++            include_dirs.append('_ctypes/darwin')
++# XXX Is this still needed?
++##            extra_link_args.extend(['-read_only_relocs', 'warning'])
++
++        elif sys.platform == 'sunos5':
++            # XXX This shouldn't be necessary; it appears that some
++            # of the assembler code is non-PIC (i.e. it has relocations
++            # when it shouldn't. The proper fix would be to rewrite
++            # the assembler code to be PIC.
++            # This only works with GCC; the Sun compiler likely refuses
++            # this option. If you want to compile ctypes with the Sun
++            # compiler, please research a proper solution, instead of
++            # finding some -z option for the Sun compiler.
++            extra_link_args.append('-mimpure-text')
++
++        elif sys.platform.startswith('hp-ux'):
++            extra_link_args.append('-fPIC')
++
++        ext = Extension('_ctypes',
++                        include_dirs=include_dirs,
++                        extra_compile_args=extra_compile_args,
++                        extra_link_args=extra_link_args,
++                        libraries=[],
++                        sources=sources,
++                        depends=depends)
++        ext_test = Extension('_ctypes_test',
++                             sources=['_ctypes/_ctypes_test.c'])
++        self.extensions.extend([ext, ext_test])
++
++        if not '--with-system-ffi' in sysconfig.get_config_var("CONFIG_ARGS"):
++            return
++
++        if sys.platform == 'darwin':
++            # OS X 10.5 comes with libffi.dylib; the include files are
++            # in /usr/include/ffi
++            inc_dirs.append('/usr/include/ffi')
++
++        ffi_inc = find_file('ffi.h', [], inc_dirs)
++        if ffi_inc is not None:
++            ffi_h = ffi_inc[0] + '/ffi.h'
++            fp = open(ffi_h)
++            while 1:
++                line = fp.readline()
++                if not line:
++                    ffi_inc = None
++                    break
++                if line.startswith('#define LIBFFI_H'):
++                    break
++        ffi_lib = None
++        if ffi_inc is not None:
++            for lib_name in ('ffi_convenience', 'ffi_pic', 'ffi'):
++                if (self.compiler.find_library_file(lib_dirs, lib_name)):
++                    ffi_lib = lib_name
++                    break
++
++        if ffi_inc and ffi_lib:
++            ext.include_dirs.extend(ffi_inc)
++            ext.libraries.append(ffi_lib)
++            self.use_system_libffi = True
++
++
++class PyBuildInstall(install):
++    # Suppress the warning about installation into the lib_dynload
++    # directory, which is not in sys.path when running Python during
++    # installation:
++    def initialize_options (self):
++        install.initialize_options(self)
++        self.warn_dir=0
++
++class PyBuildInstallLib(install_lib):
++    # Do exactly what install_lib does but make sure correct access modes get
++    # set on installed directories and files. All installed files with get
++    # mode 644 unless they are a shared library in which case they will get
++    # mode 755. All installed directories will get mode 755.
++
++    so_ext = sysconfig.get_config_var("SO")
++
++    def install(self):
++        outfiles = install_lib.install(self)
++        self.set_file_modes(outfiles, 0644, 0755)
++        self.set_dir_modes(self.install_dir, 0755)
++        return outfiles
++
++    def set_file_modes(self, files, defaultMode, sharedLibMode):
++        if not self.is_chmod_supported(): return
++        if not files: return
++
++        for filename in files:
++            if os.path.islink(filename): continue
++            mode = defaultMode
++            if filename.endswith(self.so_ext): mode = sharedLibMode
++            log.info("changing mode of %s to %o", filename, mode)
++            if not self.dry_run: os.chmod(filename, mode)
++
++    def set_dir_modes(self, dirname, mode):
++        if not self.is_chmod_supported(): return
++        os.path.walk(dirname, self.set_dir_modes_visitor, mode)
++
++    def set_dir_modes_visitor(self, mode, dirname, names):
++        if os.path.islink(dirname): return
++        log.info("changing mode of %s to %o", dirname, mode)
++        if not self.dry_run: os.chmod(dirname, mode)
++
++    def is_chmod_supported(self):
++        return hasattr(os, 'chmod')
++
++SUMMARY = """
++Python is an interpreted, interactive, object-oriented programming
++language. It is often compared to Tcl, Perl, Scheme or Java.
++
++Python combines remarkable power with very clear syntax. It has
++modules, classes, exceptions, very high level dynamic data types, and
++dynamic typing. There are interfaces to many system calls and
++libraries, as well as to various windowing systems (X11, Motif, Tk,
++Mac, MFC). New built-in modules are easily written in C or C++. Python
++is also usable as an extension language for applications that need a
++programmable interface.
++
++The Python implementation is portable: it runs on many brands of UNIX,
++on Windows, DOS, OS/2, Mac, Amiga... If your favorite system isn't
++listed here, it may still be supported, if there's a C compiler for
++it. Ask around on comp.lang.python -- or just try compiling Python
++yourself.
++"""
++
++CLASSIFIERS = """
++Development Status :: 6 - Mature
++License :: OSI Approved :: Python Software Foundation License
++Natural Language :: English
++Programming Language :: C
++Programming Language :: Python
++Topic :: Software Development
++"""
++
++def main():
++    # turn off warnings when deprecated modules are imported
++    import warnings
++    warnings.filterwarnings("ignore",category=DeprecationWarning)
++    setup(# PyPI Metadata (PEP 301)
++          name = "Python",
++          version = sys.version.split()[0],
++          url = "http://www.python.org/%s" % sys.version[:3],
++          maintainer = "Guido van Rossum and the Python community",
++          maintainer_email = "[email protected]",
++          description = "A high-level object-oriented programming language",
++          long_description = SUMMARY.strip(),
++          license = "PSF license",
++          classifiers = filter(None, CLASSIFIERS.split("\n")),
++          platforms = ["Many"],
++
++          # Build info
++          cmdclass = {'build_ext':PyBuildExt, 'install':PyBuildInstall,
++                      'install_lib':PyBuildInstallLib},
++          # The struct module is defined here, because build_ext won't be
++          # called unless there's at least one extension module defined.
++          ext_modules=[Extension('_struct', ['_struct.c'])],
++
++          # Scripts to install
++          scripts = ['Tools/scripts/pydoc', 'Tools/scripts/idle',
++                     'Tools/scripts/2to3',
++                     'Lib/smtpd.py']
++        )
++
++# --install-platlib
++if __name__ == '__main__':
++    main()
+
+diff --git Python-2.6.4/Lib/test/ucredtext.py Python-2.6.4/Lib/test/ucredtext.py
+new file mode 100644
+--- /dev/null	2011-02-12 03:14:16.000000000 -0600
++++ Python-2.6.4/Lib/test/ucredtest.py	2011-01-20 13:52:42.945657919 -0600
+@@ -0,0 +1,45 @@
++#!/usr/bin/python2.6
++
++import ucred
++import os
++
++uc = ucred.get(os.getpid())
++
++print "pid = %d" % uc.getpid()
++print "euid = %d" % uc.geteuid()
++print "ruid = %d" % uc.getruid()
++print "suid = %d" % uc.getsuid()
++print "egid = %d" % uc.getegid()
++print "rgid = %d" % uc.getrgid()
++print "sgid = %d" % uc.getsgid()
++print "zoneid = %d" % uc.getzoneid()
++print "projid = %d" % uc.getprojid()
++print "groups = %s" % uc.getgroups()
++print "label = %s" % uc.getlabel()
++
++print "getpflags(0x1) = %d" % uc.getpflags(0x1)
++print "getpflags(0x2) = %d" % uc.getpflags(0x2)
++print "has_priv(Effective, proc_fork) = %d" % uc.has_priv("Effective", "proc_fork")
++print "has_priv(Permitted, proc_fork) = %d" % uc.has_priv("Permitted", "proc_fork")
++print "has_priv(Inheritable, proc_fork) = %d" % uc.has_priv("Inheritable", "proc_fork")
++print "has_priv(Limit, file_setid) = %d" % uc.has_priv("Limit", "file_setid")
++print "has_priv(Effective, file_setid) = %d" % uc.has_priv("Effective", "file_setid")
++try:
++    uc.has_priv("Effective", "proc_bork")
++except OSError, e:
++    print e
++try:
++    uc.has_priv("Defective", "proc_fork")
++except OSError, e:
++    print e
++try:
++    uc.has_priv("Defective", "proc_bork")
++except OSError, e:
++    print e
++
++del uc
++uc = ucred.ucred()
++try:
++    uc.getpid()
++except OSError, e:
++    print e