905 lines
36 KiB
Python
905 lines
36 KiB
Python
#
|
|
# Cython Top Level
|
|
#
|
|
|
|
from __future__ import absolute_import
|
|
|
|
import os
|
|
import re
|
|
import sys
|
|
import io
|
|
|
|
if sys.version_info[:2] < (2, 6) or (3, 0) <= sys.version_info[:2] < (3, 3):
|
|
sys.stderr.write("Sorry, Cython requires Python 2.6+ or 3.3+, found %d.%d\n" % tuple(sys.version_info[:2]))
|
|
sys.exit(1)
|
|
|
|
try:
|
|
from __builtin__ import basestring
|
|
except ImportError:
|
|
basestring = str
|
|
|
|
# Do not import Parsing here, import it when needed, because Parsing imports
|
|
# Nodes, which globally needs debug command line options initialized to set a
|
|
# conditional metaclass. These options are processed by CmdLine called from
|
|
# main() in this file.
|
|
# import Parsing
|
|
from . import Errors
|
|
from .StringEncoding import EncodedString
|
|
from .Scanning import PyrexScanner, FileSourceDescriptor
|
|
from .Errors import PyrexError, CompileError, error, warning
|
|
from .Symtab import ModuleScope
|
|
from .. import Utils
|
|
from . import Options
|
|
|
|
from . import Version # legacy import needed by old PyTables versions
|
|
version = Version.version # legacy attribute - use "Cython.__version__" instead
|
|
|
|
module_name_pattern = re.compile(r"[A-Za-z_][A-Za-z0-9_]*(\.[A-Za-z_][A-Za-z0-9_]*)*$")
|
|
|
|
verbose = 0
|
|
|
|
standard_include_path = os.path.abspath(os.path.join(os.path.dirname(__file__),
|
|
os.path.pardir, 'Includes'))
|
|
|
|
class CompilationData(object):
|
|
# Bundles the information that is passed from transform to transform.
|
|
# (For now, this is only)
|
|
|
|
# While Context contains every pxd ever loaded, path information etc.,
|
|
# this only contains the data related to a single compilation pass
|
|
#
|
|
# pyx ModuleNode Main code tree of this compilation.
|
|
# pxds {string : ModuleNode} Trees for the pxds used in the pyx.
|
|
# codewriter CCodeWriter Where to output final code.
|
|
# options CompilationOptions
|
|
# result CompilationResult
|
|
pass
|
|
|
|
|
|
class Context(object):
|
|
# This class encapsulates the context needed for compiling
|
|
# one or more Cython implementation files along with their
|
|
# associated and imported declaration files. It includes
|
|
# the root of the module import namespace and the list
|
|
# of directories to search for include files.
|
|
#
|
|
# modules {string : ModuleScope}
|
|
# include_directories [string]
|
|
# future_directives [object]
|
|
# language_level int currently 2 or 3 for Python 2/3
|
|
|
|
cython_scope = None
|
|
language_level = None # warn when not set but default to Py2
|
|
|
|
def __init__(self, include_directories, compiler_directives, cpp=False,
|
|
language_level=None, options=None):
|
|
# cython_scope is a hack, set to False by subclasses, in order to break
|
|
# an infinite loop.
|
|
# Better code organization would fix it.
|
|
|
|
from . import Builtin, CythonScope
|
|
self.modules = {"__builtin__" : Builtin.builtin_scope}
|
|
self.cython_scope = CythonScope.create_cython_scope(self)
|
|
self.modules["cython"] = self.cython_scope
|
|
self.include_directories = include_directories
|
|
self.future_directives = set()
|
|
self.compiler_directives = compiler_directives
|
|
self.cpp = cpp
|
|
self.options = options
|
|
|
|
self.pxds = {} # full name -> node tree
|
|
self._interned = {} # (type(value), value, *key_args) -> interned_value
|
|
|
|
if language_level is not None:
|
|
self.set_language_level(language_level)
|
|
|
|
self.gdb_debug_outputwriter = None
|
|
|
|
def set_language_level(self, level):
|
|
from .Future import print_function, unicode_literals, absolute_import, division
|
|
future_directives = set()
|
|
if level == '3str':
|
|
level = 3
|
|
else:
|
|
level = int(level)
|
|
if level >= 3:
|
|
future_directives.add(unicode_literals)
|
|
if level >= 3:
|
|
future_directives.update([print_function, absolute_import, division])
|
|
self.language_level = level
|
|
self.future_directives = future_directives
|
|
if level >= 3:
|
|
self.modules['builtins'] = self.modules['__builtin__']
|
|
|
|
def intern_ustring(self, value, encoding=None):
|
|
key = (EncodedString, value, encoding)
|
|
try:
|
|
return self._interned[key]
|
|
except KeyError:
|
|
pass
|
|
value = EncodedString(value)
|
|
if encoding:
|
|
value.encoding = encoding
|
|
self._interned[key] = value
|
|
return value
|
|
|
|
def intern_value(self, value, *key):
|
|
key = (type(value), value) + key
|
|
try:
|
|
return self._interned[key]
|
|
except KeyError:
|
|
pass
|
|
self._interned[key] = value
|
|
return value
|
|
|
|
# pipeline creation functions can now be found in Pipeline.py
|
|
|
|
def process_pxd(self, source_desc, scope, module_name):
|
|
from . import Pipeline
|
|
if isinstance(source_desc, FileSourceDescriptor) and source_desc._file_type == 'pyx':
|
|
source = CompilationSource(source_desc, module_name, os.getcwd())
|
|
result_sink = create_default_resultobj(source, self.options)
|
|
pipeline = Pipeline.create_pyx_as_pxd_pipeline(self, result_sink)
|
|
result = Pipeline.run_pipeline(pipeline, source)
|
|
else:
|
|
pipeline = Pipeline.create_pxd_pipeline(self, scope, module_name)
|
|
result = Pipeline.run_pipeline(pipeline, source_desc)
|
|
return result
|
|
|
|
def nonfatal_error(self, exc):
|
|
return Errors.report_error(exc)
|
|
|
|
def find_module(self, module_name, relative_to=None, pos=None, need_pxd=1,
|
|
absolute_fallback=True):
|
|
# Finds and returns the module scope corresponding to
|
|
# the given relative or absolute module name. If this
|
|
# is the first time the module has been requested, finds
|
|
# the corresponding .pxd file and process it.
|
|
# If relative_to is not None, it must be a module scope,
|
|
# and the module will first be searched for relative to
|
|
# that module, provided its name is not a dotted name.
|
|
debug_find_module = 0
|
|
if debug_find_module:
|
|
print("Context.find_module: module_name = %s, relative_to = %s, pos = %s, need_pxd = %s" % (
|
|
module_name, relative_to, pos, need_pxd))
|
|
|
|
scope = None
|
|
pxd_pathname = None
|
|
if relative_to:
|
|
if module_name:
|
|
# from .module import ...
|
|
qualified_name = relative_to.qualify_name(module_name)
|
|
else:
|
|
# from . import ...
|
|
qualified_name = relative_to.qualified_name
|
|
scope = relative_to
|
|
relative_to = None
|
|
else:
|
|
qualified_name = module_name
|
|
|
|
if not module_name_pattern.match(qualified_name):
|
|
raise CompileError(pos or (module_name, 0, 0),
|
|
"'%s' is not a valid module name" % module_name)
|
|
|
|
if relative_to:
|
|
if debug_find_module:
|
|
print("...trying relative import")
|
|
scope = relative_to.lookup_submodule(module_name)
|
|
if not scope:
|
|
pxd_pathname = self.find_pxd_file(qualified_name, pos)
|
|
if pxd_pathname:
|
|
scope = relative_to.find_submodule(module_name)
|
|
if not scope:
|
|
if debug_find_module:
|
|
print("...trying absolute import")
|
|
if absolute_fallback:
|
|
qualified_name = module_name
|
|
scope = self
|
|
for name in qualified_name.split("."):
|
|
scope = scope.find_submodule(name)
|
|
|
|
if debug_find_module:
|
|
print("...scope = %s" % scope)
|
|
if not scope.pxd_file_loaded:
|
|
if debug_find_module:
|
|
print("...pxd not loaded")
|
|
if not pxd_pathname:
|
|
if debug_find_module:
|
|
print("...looking for pxd file")
|
|
# Only look in sys.path if we are explicitly looking
|
|
# for a .pxd file.
|
|
pxd_pathname = self.find_pxd_file(qualified_name, pos, sys_path=need_pxd)
|
|
if debug_find_module:
|
|
print("......found %s" % pxd_pathname)
|
|
if not pxd_pathname and need_pxd:
|
|
# Set pxd_file_loaded such that we don't need to
|
|
# look for the non-existing pxd file next time.
|
|
scope.pxd_file_loaded = True
|
|
package_pathname = self.search_include_directories(qualified_name, ".py", pos)
|
|
if package_pathname and package_pathname.endswith('__init__.py'):
|
|
pass
|
|
else:
|
|
error(pos, "'%s.pxd' not found" % qualified_name.replace('.', os.sep))
|
|
if pxd_pathname:
|
|
scope.pxd_file_loaded = True
|
|
try:
|
|
if debug_find_module:
|
|
print("Context.find_module: Parsing %s" % pxd_pathname)
|
|
rel_path = module_name.replace('.', os.sep) + os.path.splitext(pxd_pathname)[1]
|
|
if not pxd_pathname.endswith(rel_path):
|
|
rel_path = pxd_pathname # safety measure to prevent printing incorrect paths
|
|
source_desc = FileSourceDescriptor(pxd_pathname, rel_path)
|
|
err, result = self.process_pxd(source_desc, scope, qualified_name)
|
|
if err:
|
|
raise err
|
|
(pxd_codenodes, pxd_scope) = result
|
|
self.pxds[module_name] = (pxd_codenodes, pxd_scope)
|
|
except CompileError:
|
|
pass
|
|
return scope
|
|
|
|
def find_pxd_file(self, qualified_name, pos, sys_path=True):
|
|
# Search include path (and sys.path if sys_path is True) for
|
|
# the .pxd file corresponding to the given fully-qualified
|
|
# module name.
|
|
# Will find either a dotted filename or a file in a
|
|
# package directory. If a source file position is given,
|
|
# the directory containing the source file is searched first
|
|
# for a dotted filename, and its containing package root
|
|
# directory is searched first for a non-dotted filename.
|
|
pxd = self.search_include_directories(qualified_name, ".pxd", pos, sys_path=sys_path)
|
|
if pxd is None: # XXX Keep this until Includes/Deprecated is removed
|
|
if (qualified_name.startswith('python') or
|
|
qualified_name in ('stdlib', 'stdio', 'stl')):
|
|
standard_include_path = os.path.abspath(os.path.normpath(
|
|
os.path.join(os.path.dirname(__file__), os.path.pardir, 'Includes')))
|
|
deprecated_include_path = os.path.join(standard_include_path, 'Deprecated')
|
|
self.include_directories.append(deprecated_include_path)
|
|
try:
|
|
pxd = self.search_include_directories(qualified_name, ".pxd", pos)
|
|
finally:
|
|
self.include_directories.pop()
|
|
if pxd:
|
|
name = qualified_name
|
|
if name.startswith('python'):
|
|
warning(pos, "'%s' is deprecated, use 'cpython'" % name, 1)
|
|
elif name in ('stdlib', 'stdio'):
|
|
warning(pos, "'%s' is deprecated, use 'libc.%s'" % (name, name), 1)
|
|
elif name in ('stl'):
|
|
warning(pos, "'%s' is deprecated, use 'libcpp.*.*'" % name, 1)
|
|
if pxd is None and Options.cimport_from_pyx:
|
|
return self.find_pyx_file(qualified_name, pos)
|
|
return pxd
|
|
|
|
def find_pyx_file(self, qualified_name, pos):
|
|
# Search include path for the .pyx file corresponding to the
|
|
# given fully-qualified module name, as for find_pxd_file().
|
|
return self.search_include_directories(qualified_name, ".pyx", pos)
|
|
|
|
def find_include_file(self, filename, pos):
|
|
# Search list of include directories for filename.
|
|
# Reports an error and returns None if not found.
|
|
path = self.search_include_directories(filename, "", pos,
|
|
include=True)
|
|
if not path:
|
|
error(pos, "'%s' not found" % filename)
|
|
return path
|
|
|
|
def search_include_directories(self, qualified_name, suffix, pos,
|
|
include=False, sys_path=False):
|
|
include_dirs = self.include_directories
|
|
if sys_path:
|
|
include_dirs = include_dirs + sys.path
|
|
# include_dirs must be hashable for caching in @cached_function
|
|
include_dirs = tuple(include_dirs + [standard_include_path])
|
|
return search_include_directories(include_dirs, qualified_name,
|
|
suffix, pos, include)
|
|
|
|
def find_root_package_dir(self, file_path):
|
|
return Utils.find_root_package_dir(file_path)
|
|
|
|
def check_package_dir(self, dir, package_names):
|
|
return Utils.check_package_dir(dir, tuple(package_names))
|
|
|
|
def c_file_out_of_date(self, source_path, output_path):
|
|
if not os.path.exists(output_path):
|
|
return 1
|
|
c_time = Utils.modification_time(output_path)
|
|
if Utils.file_newer_than(source_path, c_time):
|
|
return 1
|
|
pos = [source_path]
|
|
pxd_path = Utils.replace_suffix(source_path, ".pxd")
|
|
if os.path.exists(pxd_path) and Utils.file_newer_than(pxd_path, c_time):
|
|
return 1
|
|
for kind, name in self.read_dependency_file(source_path):
|
|
if kind == "cimport":
|
|
dep_path = self.find_pxd_file(name, pos)
|
|
elif kind == "include":
|
|
dep_path = self.search_include_directories(name, pos)
|
|
else:
|
|
continue
|
|
if dep_path and Utils.file_newer_than(dep_path, c_time):
|
|
return 1
|
|
return 0
|
|
|
|
def find_cimported_module_names(self, source_path):
|
|
return [ name for kind, name in self.read_dependency_file(source_path)
|
|
if kind == "cimport" ]
|
|
|
|
def is_package_dir(self, dir_path):
|
|
return Utils.is_package_dir(dir_path)
|
|
|
|
def read_dependency_file(self, source_path):
|
|
dep_path = Utils.replace_suffix(source_path, ".dep")
|
|
if os.path.exists(dep_path):
|
|
f = open(dep_path, "rU")
|
|
chunks = [ line.strip().split(" ", 1)
|
|
for line in f.readlines()
|
|
if " " in line.strip() ]
|
|
f.close()
|
|
return chunks
|
|
else:
|
|
return ()
|
|
|
|
def lookup_submodule(self, name):
|
|
# Look up a top-level module. Returns None if not found.
|
|
return self.modules.get(name, None)
|
|
|
|
def find_submodule(self, name):
|
|
# Find a top-level module, creating a new one if needed.
|
|
scope = self.lookup_submodule(name)
|
|
if not scope:
|
|
scope = ModuleScope(name,
|
|
parent_module = None, context = self)
|
|
self.modules[name] = scope
|
|
return scope
|
|
|
|
def parse(self, source_desc, scope, pxd, full_module_name):
|
|
if not isinstance(source_desc, FileSourceDescriptor):
|
|
raise RuntimeError("Only file sources for code supported")
|
|
source_filename = source_desc.filename
|
|
scope.cpp = self.cpp
|
|
# Parse the given source file and return a parse tree.
|
|
num_errors = Errors.num_errors
|
|
try:
|
|
with Utils.open_source_file(source_filename) as f:
|
|
from . import Parsing
|
|
s = PyrexScanner(f, source_desc, source_encoding = f.encoding,
|
|
scope = scope, context = self)
|
|
tree = Parsing.p_module(s, pxd, full_module_name)
|
|
if self.options.formal_grammar:
|
|
try:
|
|
from ..Parser import ConcreteSyntaxTree
|
|
except ImportError:
|
|
raise RuntimeError(
|
|
"Formal grammar can only be used with compiled Cython with an available pgen.")
|
|
ConcreteSyntaxTree.p_module(source_filename)
|
|
except UnicodeDecodeError as e:
|
|
#import traceback
|
|
#traceback.print_exc()
|
|
raise self._report_decode_error(source_desc, e)
|
|
|
|
if Errors.num_errors > num_errors:
|
|
raise CompileError()
|
|
return tree
|
|
|
|
def _report_decode_error(self, source_desc, exc):
|
|
msg = exc.args[-1]
|
|
position = exc.args[2]
|
|
encoding = exc.args[0]
|
|
|
|
line = 1
|
|
column = idx = 0
|
|
with io.open(source_desc.filename, "r", encoding='iso8859-1', newline='') as f:
|
|
for line, data in enumerate(f, 1):
|
|
idx += len(data)
|
|
if idx >= position:
|
|
column = position - (idx - len(data)) + 1
|
|
break
|
|
|
|
return error((source_desc, line, column),
|
|
"Decoding error, missing or incorrect coding=<encoding-name> "
|
|
"at top of source (cannot decode with encoding %r: %s)" % (encoding, msg))
|
|
|
|
def extract_module_name(self, path, options):
|
|
# Find fully_qualified module name from the full pathname
|
|
# of a source file.
|
|
dir, filename = os.path.split(path)
|
|
module_name, _ = os.path.splitext(filename)
|
|
if "." in module_name:
|
|
return module_name
|
|
names = [module_name]
|
|
while self.is_package_dir(dir):
|
|
parent, package_name = os.path.split(dir)
|
|
if parent == dir:
|
|
break
|
|
names.append(package_name)
|
|
dir = parent
|
|
names.reverse()
|
|
return ".".join(names)
|
|
|
|
def setup_errors(self, options, result):
|
|
Errors.reset() # clear any remaining error state
|
|
if options.use_listing_file:
|
|
path = result.listing_file = Utils.replace_suffix(result.main_source_file, ".lis")
|
|
else:
|
|
path = None
|
|
Errors.open_listing_file(path=path,
|
|
echo_to_stderr=options.errors_to_stderr)
|
|
|
|
def teardown_errors(self, err, options, result):
|
|
source_desc = result.compilation_source.source_desc
|
|
if not isinstance(source_desc, FileSourceDescriptor):
|
|
raise RuntimeError("Only file sources for code supported")
|
|
Errors.close_listing_file()
|
|
result.num_errors = Errors.num_errors
|
|
if result.num_errors > 0:
|
|
err = True
|
|
if err and result.c_file:
|
|
try:
|
|
Utils.castrate_file(result.c_file, os.stat(source_desc.filename))
|
|
except EnvironmentError:
|
|
pass
|
|
result.c_file = None
|
|
|
|
|
|
def get_output_filename(source_filename, cwd, options):
|
|
if options.cplus:
|
|
c_suffix = ".cpp"
|
|
else:
|
|
c_suffix = ".c"
|
|
suggested_file_name = Utils.replace_suffix(source_filename, c_suffix)
|
|
if options.output_file:
|
|
out_path = os.path.join(cwd, options.output_file)
|
|
if os.path.isdir(out_path):
|
|
return os.path.join(out_path, os.path.basename(suggested_file_name))
|
|
else:
|
|
return out_path
|
|
else:
|
|
return suggested_file_name
|
|
|
|
|
|
def create_default_resultobj(compilation_source, options):
|
|
result = CompilationResult()
|
|
result.main_source_file = compilation_source.source_desc.filename
|
|
result.compilation_source = compilation_source
|
|
source_desc = compilation_source.source_desc
|
|
result.c_file = get_output_filename(source_desc.filename,
|
|
compilation_source.cwd, options)
|
|
result.embedded_metadata = options.embedded_metadata
|
|
return result
|
|
|
|
|
|
def run_pipeline(source, options, full_module_name=None, context=None):
|
|
from . import Pipeline
|
|
|
|
source_ext = os.path.splitext(source)[1]
|
|
options.configure_language_defaults(source_ext[1:]) # py/pyx
|
|
if context is None:
|
|
context = options.create_context()
|
|
|
|
# Set up source object
|
|
cwd = os.getcwd()
|
|
abs_path = os.path.abspath(source)
|
|
full_module_name = full_module_name or context.extract_module_name(source, options)
|
|
|
|
Utils.raise_error_if_module_name_forbidden(full_module_name)
|
|
|
|
if options.relative_path_in_code_position_comments:
|
|
rel_path = full_module_name.replace('.', os.sep) + source_ext
|
|
if not abs_path.endswith(rel_path):
|
|
rel_path = source # safety measure to prevent printing incorrect paths
|
|
else:
|
|
rel_path = abs_path
|
|
source_desc = FileSourceDescriptor(abs_path, rel_path)
|
|
source = CompilationSource(source_desc, full_module_name, cwd)
|
|
|
|
# Set up result object
|
|
result = create_default_resultobj(source, options)
|
|
|
|
if options.annotate is None:
|
|
# By default, decide based on whether an html file already exists.
|
|
html_filename = os.path.splitext(result.c_file)[0] + ".html"
|
|
if os.path.exists(html_filename):
|
|
with io.open(html_filename, "r", encoding="UTF-8") as html_file:
|
|
if u'<!-- Generated by Cython' in html_file.read(100):
|
|
options.annotate = True
|
|
|
|
# Get pipeline
|
|
if source_ext.lower() == '.py' or not source_ext:
|
|
pipeline = Pipeline.create_py_pipeline(context, options, result)
|
|
else:
|
|
pipeline = Pipeline.create_pyx_pipeline(context, options, result)
|
|
|
|
context.setup_errors(options, result)
|
|
err, enddata = Pipeline.run_pipeline(pipeline, source)
|
|
context.teardown_errors(err, options, result)
|
|
return result
|
|
|
|
|
|
# ------------------------------------------------------------------------
|
|
#
|
|
# Main Python entry points
|
|
#
|
|
# ------------------------------------------------------------------------
|
|
|
|
class CompilationSource(object):
|
|
"""
|
|
Contains the data necessary to start up a compilation pipeline for
|
|
a single compilation unit.
|
|
"""
|
|
def __init__(self, source_desc, full_module_name, cwd):
|
|
self.source_desc = source_desc
|
|
self.full_module_name = full_module_name
|
|
self.cwd = cwd
|
|
|
|
|
|
class CompilationOptions(object):
|
|
r"""
|
|
See default_options at the end of this module for a list of all possible
|
|
options and CmdLine.usage and CmdLine.parse_command_line() for their
|
|
meaning.
|
|
"""
|
|
def __init__(self, defaults=None, **kw):
|
|
self.include_path = []
|
|
if defaults:
|
|
if isinstance(defaults, CompilationOptions):
|
|
defaults = defaults.__dict__
|
|
else:
|
|
defaults = default_options
|
|
|
|
options = dict(defaults)
|
|
options.update(kw)
|
|
|
|
# let's assume 'default_options' contains a value for most known compiler options
|
|
# and validate against them
|
|
unknown_options = set(options) - set(default_options)
|
|
# ignore valid options that are not in the defaults
|
|
unknown_options.difference_update(['include_path'])
|
|
if unknown_options:
|
|
message = "got unknown compilation option%s, please remove: %s" % (
|
|
's' if len(unknown_options) > 1 else '',
|
|
', '.join(unknown_options))
|
|
raise ValueError(message)
|
|
|
|
directive_defaults = Options.get_directive_defaults()
|
|
directives = dict(options['compiler_directives']) # copy mutable field
|
|
# check for invalid directives
|
|
unknown_directives = set(directives) - set(directive_defaults)
|
|
if unknown_directives:
|
|
message = "got unknown compiler directive%s: %s" % (
|
|
's' if len(unknown_directives) > 1 else '',
|
|
', '.join(unknown_directives))
|
|
raise ValueError(message)
|
|
options['compiler_directives'] = directives
|
|
if directives.get('np_pythran', False) and not options['cplus']:
|
|
import warnings
|
|
warnings.warn("C++ mode forced when in Pythran mode!")
|
|
options['cplus'] = True
|
|
if 'language_level' in directives and 'language_level' not in kw:
|
|
options['language_level'] = directives['language_level']
|
|
elif not options.get('language_level'):
|
|
options['language_level'] = directive_defaults.get('language_level')
|
|
if 'formal_grammar' in directives and 'formal_grammar' not in kw:
|
|
options['formal_grammar'] = directives['formal_grammar']
|
|
if options['cache'] is True:
|
|
options['cache'] = os.path.join(Utils.get_cython_cache_dir(), 'compiler')
|
|
|
|
self.__dict__.update(options)
|
|
|
|
def configure_language_defaults(self, source_extension):
|
|
if source_extension == 'py':
|
|
if self.compiler_directives.get('binding') is None:
|
|
self.compiler_directives['binding'] = True
|
|
|
|
def create_context(self):
|
|
return Context(self.include_path, self.compiler_directives,
|
|
self.cplus, self.language_level, options=self)
|
|
|
|
def get_fingerprint(self):
|
|
r"""
|
|
Return a string that contains all the options that are relevant for cache invalidation.
|
|
"""
|
|
# Collect only the data that can affect the generated file(s).
|
|
data = {}
|
|
|
|
for key, value in self.__dict__.items():
|
|
if key in ['show_version', 'errors_to_stderr', 'verbose', 'quiet']:
|
|
# verbosity flags have no influence on the compilation result
|
|
continue
|
|
elif key in ['output_file', 'output_dir']:
|
|
# ignore the exact name of the output file
|
|
continue
|
|
elif key in ['timestamps']:
|
|
# the cache cares about the content of files, not about the timestamps of sources
|
|
continue
|
|
elif key in ['cache']:
|
|
# hopefully caching has no influence on the compilation result
|
|
continue
|
|
elif key in ['compiler_directives']:
|
|
# directives passed on to the C compiler do not influence the generated C code
|
|
continue
|
|
elif key in ['include_path']:
|
|
# this path changes which headers are tracked as dependencies,
|
|
# it has no influence on the generated C code
|
|
continue
|
|
elif key in ['working_path']:
|
|
# this path changes where modules and pxd files are found;
|
|
# their content is part of the fingerprint anyway, their
|
|
# absolute path does not matter
|
|
continue
|
|
elif key in ['create_extension']:
|
|
# create_extension() has already mangled the options, e.g.,
|
|
# embedded_metadata, when the fingerprint is computed so we
|
|
# ignore it here.
|
|
continue
|
|
elif key in ['build_dir']:
|
|
# the (temporary) directory where we collect dependencies
|
|
# has no influence on the C output
|
|
continue
|
|
elif key in ['use_listing_file', 'generate_pxi', 'annotate', 'annotate_coverage_xml']:
|
|
# all output files are contained in the cache so the types of
|
|
# files generated must be part of the fingerprint
|
|
data[key] = value
|
|
elif key in ['formal_grammar', 'evaluate_tree_assertions']:
|
|
# these bits can change whether compilation to C passes/fails
|
|
data[key] = value
|
|
elif key in ['embedded_metadata', 'emit_linenums', 'c_line_in_traceback', 'gdb_debug', 'relative_path_in_code_position_comments']:
|
|
# the generated code contains additional bits when these are set
|
|
data[key] = value
|
|
elif key in ['cplus', 'language_level', 'compile_time_env', 'np_pythran']:
|
|
# assorted bits that, e.g., influence the parser
|
|
data[key] = value
|
|
elif key == ['capi_reexport_cincludes']:
|
|
if self.capi_reexport_cincludes:
|
|
# our caching implementation does not yet include fingerprints of all the header files
|
|
raise NotImplementedError('capi_reexport_cincludes is not compatible with Cython caching')
|
|
elif key == ['common_utility_include_dir']:
|
|
if self.common_utility_include_dir:
|
|
raise NotImplementedError('common_utility_include_dir is not compatible with Cython caching yet')
|
|
else:
|
|
# any unexpected option should go into the fingerprint; it's better
|
|
# to recompile than to return incorrect results from the cache.
|
|
data[key] = value
|
|
|
|
def to_fingerprint(item):
|
|
r"""
|
|
Recursively turn item into a string, turning dicts into lists with
|
|
deterministic ordering.
|
|
"""
|
|
if isinstance(item, dict):
|
|
item = sorted([(repr(key), to_fingerprint(value)) for key, value in item.items()])
|
|
return repr(item)
|
|
|
|
return to_fingerprint(data)
|
|
|
|
|
|
class CompilationResult(object):
|
|
"""
|
|
Results from the Cython compiler:
|
|
|
|
c_file string or None The generated C source file
|
|
h_file string or None The generated C header file
|
|
i_file string or None The generated .pxi file
|
|
api_file string or None The generated C API .h file
|
|
listing_file string or None File of error messages
|
|
object_file string or None Result of compiling the C file
|
|
extension_file string or None Result of linking the object file
|
|
num_errors integer Number of compilation errors
|
|
compilation_source CompilationSource
|
|
"""
|
|
|
|
def __init__(self):
|
|
self.c_file = None
|
|
self.h_file = None
|
|
self.i_file = None
|
|
self.api_file = None
|
|
self.listing_file = None
|
|
self.object_file = None
|
|
self.extension_file = None
|
|
self.main_source_file = None
|
|
|
|
|
|
class CompilationResultSet(dict):
|
|
"""
|
|
Results from compiling multiple Pyrex source files. A mapping
|
|
from source file paths to CompilationResult instances. Also
|
|
has the following attributes:
|
|
|
|
num_errors integer Total number of compilation errors
|
|
"""
|
|
|
|
num_errors = 0
|
|
|
|
def add(self, source, result):
|
|
self[source] = result
|
|
self.num_errors += result.num_errors
|
|
|
|
|
|
def compile_single(source, options, full_module_name = None):
|
|
"""
|
|
compile_single(source, options, full_module_name)
|
|
|
|
Compile the given Pyrex implementation file and return a CompilationResult.
|
|
Always compiles a single file; does not perform timestamp checking or
|
|
recursion.
|
|
"""
|
|
return run_pipeline(source, options, full_module_name)
|
|
|
|
|
|
def compile_multiple(sources, options):
|
|
"""
|
|
compile_multiple(sources, options)
|
|
|
|
Compiles the given sequence of Pyrex implementation files and returns
|
|
a CompilationResultSet. Performs timestamp checking and/or recursion
|
|
if these are specified in the options.
|
|
"""
|
|
# run_pipeline creates the context
|
|
# context = options.create_context()
|
|
sources = [os.path.abspath(source) for source in sources]
|
|
processed = set()
|
|
results = CompilationResultSet()
|
|
timestamps = options.timestamps
|
|
verbose = options.verbose
|
|
context = None
|
|
cwd = os.getcwd()
|
|
for source in sources:
|
|
if source not in processed:
|
|
if context is None:
|
|
context = options.create_context()
|
|
output_filename = get_output_filename(source, cwd, options)
|
|
out_of_date = context.c_file_out_of_date(source, output_filename)
|
|
if (not timestamps) or out_of_date:
|
|
if verbose:
|
|
sys.stderr.write("Compiling %s\n" % source)
|
|
|
|
result = run_pipeline(source, options, context=context)
|
|
results.add(source, result)
|
|
# Compiling multiple sources in one context doesn't quite
|
|
# work properly yet.
|
|
context = None
|
|
processed.add(source)
|
|
return results
|
|
|
|
|
|
def compile(source, options = None, full_module_name = None, **kwds):
|
|
"""
|
|
compile(source [, options], [, <option> = <value>]...)
|
|
|
|
Compile one or more Pyrex implementation files, with optional timestamp
|
|
checking and recursing on dependencies. The source argument may be a string
|
|
or a sequence of strings. If it is a string and no recursion or timestamp
|
|
checking is requested, a CompilationResult is returned, otherwise a
|
|
CompilationResultSet is returned.
|
|
"""
|
|
options = CompilationOptions(defaults = options, **kwds)
|
|
if isinstance(source, basestring) and not options.timestamps:
|
|
return compile_single(source, options, full_module_name)
|
|
else:
|
|
return compile_multiple(source, options)
|
|
|
|
|
|
@Utils.cached_function
|
|
def search_include_directories(dirs, qualified_name, suffix, pos, include=False):
|
|
"""
|
|
Search the list of include directories for the given file name.
|
|
|
|
If a source file position is given, first searches the directory
|
|
containing that file. Returns None if not found, but does not
|
|
report an error.
|
|
|
|
The 'include' option will disable package dereferencing.
|
|
"""
|
|
|
|
if pos:
|
|
file_desc = pos[0]
|
|
if not isinstance(file_desc, FileSourceDescriptor):
|
|
raise RuntimeError("Only file sources for code supported")
|
|
if include:
|
|
dirs = (os.path.dirname(file_desc.filename),) + dirs
|
|
else:
|
|
dirs = (Utils.find_root_package_dir(file_desc.filename),) + dirs
|
|
|
|
dotted_filename = qualified_name
|
|
if suffix:
|
|
dotted_filename += suffix
|
|
|
|
if not include:
|
|
names = qualified_name.split('.')
|
|
package_names = tuple(names[:-1])
|
|
module_name = names[-1]
|
|
module_filename = module_name + suffix
|
|
package_filename = "__init__" + suffix
|
|
|
|
for dirname in dirs:
|
|
path = os.path.join(dirname, dotted_filename)
|
|
if os.path.exists(path):
|
|
return path
|
|
|
|
if not include:
|
|
package_dir = Utils.check_package_dir(dirname, package_names)
|
|
if package_dir is not None:
|
|
path = os.path.join(package_dir, module_filename)
|
|
if os.path.exists(path):
|
|
return path
|
|
path = os.path.join(package_dir, module_name,
|
|
package_filename)
|
|
if os.path.exists(path):
|
|
return path
|
|
return None
|
|
|
|
|
|
# ------------------------------------------------------------------------
|
|
#
|
|
# Main command-line entry point
|
|
#
|
|
# ------------------------------------------------------------------------
|
|
|
|
def setuptools_main():
|
|
return main(command_line = 1)
|
|
|
|
|
|
def main(command_line = 0):
|
|
args = sys.argv[1:]
|
|
any_failures = 0
|
|
if command_line:
|
|
from .CmdLine import parse_command_line
|
|
options, sources = parse_command_line(args)
|
|
else:
|
|
options = CompilationOptions(default_options)
|
|
sources = args
|
|
|
|
if options.show_version:
|
|
sys.stderr.write("Cython version %s\n" % version)
|
|
if options.working_path!="":
|
|
os.chdir(options.working_path)
|
|
try:
|
|
result = compile(sources, options)
|
|
if result.num_errors > 0:
|
|
any_failures = 1
|
|
except (EnvironmentError, PyrexError) as e:
|
|
sys.stderr.write(str(e) + '\n')
|
|
any_failures = 1
|
|
if any_failures:
|
|
sys.exit(1)
|
|
|
|
|
|
# ------------------------------------------------------------------------
|
|
#
|
|
# Set the default options depending on the platform
|
|
#
|
|
# ------------------------------------------------------------------------
|
|
|
|
default_options = dict(
|
|
show_version = 0,
|
|
use_listing_file = 0,
|
|
errors_to_stderr = 1,
|
|
cplus = 0,
|
|
output_file = None,
|
|
annotate = None,
|
|
annotate_coverage_xml = None,
|
|
generate_pxi = 0,
|
|
capi_reexport_cincludes = 0,
|
|
working_path = "",
|
|
timestamps = None,
|
|
verbose = 0,
|
|
quiet = 0,
|
|
compiler_directives = {},
|
|
embedded_metadata = {},
|
|
evaluate_tree_assertions = False,
|
|
emit_linenums = False,
|
|
relative_path_in_code_position_comments = True,
|
|
c_line_in_traceback = True,
|
|
language_level = None, # warn but default to 2
|
|
formal_grammar = False,
|
|
gdb_debug = False,
|
|
compile_time_env = None,
|
|
common_utility_include_dir = None,
|
|
output_dir=None,
|
|
build_dir=None,
|
|
cache=None,
|
|
create_extension=None,
|
|
np_pythran=False
|
|
)
|