1494 lines
48 KiB
Cython
1494 lines
48 KiB
Cython
|
#################### View.MemoryView ####################
|
||
|
|
||
|
# This utility provides cython.array and cython.view.memoryview
|
||
|
|
||
|
from __future__ import absolute_import
|
||
|
|
||
|
cimport cython
|
||
|
|
||
|
# from cpython cimport ...
|
||
|
cdef extern from "Python.h":
|
||
|
int PyIndex_Check(object)
|
||
|
object PyLong_FromVoidPtr(void *)
|
||
|
|
||
|
cdef extern from "pythread.h":
|
||
|
ctypedef void *PyThread_type_lock
|
||
|
|
||
|
PyThread_type_lock PyThread_allocate_lock()
|
||
|
void PyThread_free_lock(PyThread_type_lock)
|
||
|
int PyThread_acquire_lock(PyThread_type_lock, int mode) nogil
|
||
|
void PyThread_release_lock(PyThread_type_lock) nogil
|
||
|
|
||
|
cdef extern from "<string.h>":
|
||
|
void *memset(void *b, int c, size_t len)
|
||
|
|
||
|
cdef extern from *:
|
||
|
int __Pyx_GetBuffer(object, Py_buffer *, int) except -1
|
||
|
void __Pyx_ReleaseBuffer(Py_buffer *)
|
||
|
|
||
|
ctypedef struct PyObject
|
||
|
ctypedef Py_ssize_t Py_intptr_t
|
||
|
void Py_INCREF(PyObject *)
|
||
|
void Py_DECREF(PyObject *)
|
||
|
|
||
|
void* PyMem_Malloc(size_t n)
|
||
|
void PyMem_Free(void *p)
|
||
|
void* PyObject_Malloc(size_t n)
|
||
|
void PyObject_Free(void *p)
|
||
|
|
||
|
cdef struct __pyx_memoryview "__pyx_memoryview_obj":
|
||
|
Py_buffer view
|
||
|
PyObject *obj
|
||
|
__Pyx_TypeInfo *typeinfo
|
||
|
|
||
|
ctypedef struct {{memviewslice_name}}:
|
||
|
__pyx_memoryview *memview
|
||
|
char *data
|
||
|
Py_ssize_t shape[{{max_dims}}]
|
||
|
Py_ssize_t strides[{{max_dims}}]
|
||
|
Py_ssize_t suboffsets[{{max_dims}}]
|
||
|
|
||
|
void __PYX_INC_MEMVIEW({{memviewslice_name}} *memslice, int have_gil)
|
||
|
void __PYX_XDEC_MEMVIEW({{memviewslice_name}} *memslice, int have_gil)
|
||
|
|
||
|
ctypedef struct __pyx_buffer "Py_buffer":
|
||
|
PyObject *obj
|
||
|
|
||
|
PyObject *Py_None
|
||
|
|
||
|
cdef enum:
|
||
|
PyBUF_C_CONTIGUOUS,
|
||
|
PyBUF_F_CONTIGUOUS,
|
||
|
PyBUF_ANY_CONTIGUOUS
|
||
|
PyBUF_FORMAT
|
||
|
PyBUF_WRITABLE
|
||
|
PyBUF_STRIDES
|
||
|
PyBUF_INDIRECT
|
||
|
PyBUF_ND
|
||
|
PyBUF_RECORDS
|
||
|
PyBUF_RECORDS_RO
|
||
|
|
||
|
ctypedef struct __Pyx_TypeInfo:
|
||
|
pass
|
||
|
|
||
|
cdef object capsule "__pyx_capsule_create" (void *p, char *sig)
|
||
|
cdef int __pyx_array_getbuffer(PyObject *obj, Py_buffer view, int flags)
|
||
|
cdef int __pyx_memoryview_getbuffer(PyObject *obj, Py_buffer view, int flags)
|
||
|
|
||
|
cdef extern from *:
|
||
|
ctypedef int __pyx_atomic_int
|
||
|
{{memviewslice_name}} slice_copy_contig "__pyx_memoryview_copy_new_contig"(
|
||
|
__Pyx_memviewslice *from_mvs,
|
||
|
char *mode, int ndim,
|
||
|
size_t sizeof_dtype, int contig_flag,
|
||
|
bint dtype_is_object) nogil except *
|
||
|
bint slice_is_contig "__pyx_memviewslice_is_contig" (
|
||
|
{{memviewslice_name}} mvs, char order, int ndim) nogil
|
||
|
bint slices_overlap "__pyx_slices_overlap" ({{memviewslice_name}} *slice1,
|
||
|
{{memviewslice_name}} *slice2,
|
||
|
int ndim, size_t itemsize) nogil
|
||
|
|
||
|
|
||
|
cdef extern from "<stdlib.h>":
|
||
|
void *malloc(size_t) nogil
|
||
|
void free(void *) nogil
|
||
|
void *memcpy(void *dest, void *src, size_t n) nogil
|
||
|
|
||
|
|
||
|
|
||
|
|
||
|
#
|
||
|
### cython.array class
|
||
|
#
|
||
|
|
||
|
@cname("__pyx_array")
|
||
|
cdef class array:
|
||
|
|
||
|
cdef:
|
||
|
char *data
|
||
|
Py_ssize_t len
|
||
|
char *format
|
||
|
int ndim
|
||
|
Py_ssize_t *_shape
|
||
|
Py_ssize_t *_strides
|
||
|
Py_ssize_t itemsize
|
||
|
unicode mode # FIXME: this should have been a simple 'char'
|
||
|
bytes _format
|
||
|
void (*callback_free_data)(void *data)
|
||
|
# cdef object _memview
|
||
|
cdef bint free_data
|
||
|
cdef bint dtype_is_object
|
||
|
|
||
|
def __cinit__(array self, tuple shape, Py_ssize_t itemsize, format not None,
|
||
|
mode="c", bint allocate_buffer=True):
|
||
|
|
||
|
cdef int idx
|
||
|
cdef Py_ssize_t i, dim
|
||
|
cdef PyObject **p
|
||
|
|
||
|
self.ndim = <int> len(shape)
|
||
|
self.itemsize = itemsize
|
||
|
|
||
|
if not self.ndim:
|
||
|
raise ValueError("Empty shape tuple for cython.array")
|
||
|
|
||
|
if itemsize <= 0:
|
||
|
raise ValueError("itemsize <= 0 for cython.array")
|
||
|
|
||
|
if not isinstance(format, bytes):
|
||
|
format = format.encode('ASCII')
|
||
|
self._format = format # keep a reference to the byte string
|
||
|
self.format = self._format
|
||
|
|
||
|
# use single malloc() for both shape and strides
|
||
|
self._shape = <Py_ssize_t *> PyObject_Malloc(sizeof(Py_ssize_t)*self.ndim*2)
|
||
|
self._strides = self._shape + self.ndim
|
||
|
|
||
|
if not self._shape:
|
||
|
raise MemoryError("unable to allocate shape and strides.")
|
||
|
|
||
|
# cdef Py_ssize_t dim, stride
|
||
|
for idx, dim in enumerate(shape):
|
||
|
if dim <= 0:
|
||
|
raise ValueError("Invalid shape in axis %d: %d." % (idx, dim))
|
||
|
self._shape[idx] = dim
|
||
|
|
||
|
cdef char order
|
||
|
if mode == 'fortran':
|
||
|
order = b'F'
|
||
|
self.mode = u'fortran'
|
||
|
elif mode == 'c':
|
||
|
order = b'C'
|
||
|
self.mode = u'c'
|
||
|
else:
|
||
|
raise ValueError("Invalid mode, expected 'c' or 'fortran', got %s" % mode)
|
||
|
|
||
|
self.len = fill_contig_strides_array(self._shape, self._strides,
|
||
|
itemsize, self.ndim, order)
|
||
|
|
||
|
self.free_data = allocate_buffer
|
||
|
self.dtype_is_object = format == b'O'
|
||
|
if allocate_buffer:
|
||
|
# use malloc() for backwards compatibility
|
||
|
# in case external code wants to change the data pointer
|
||
|
self.data = <char *>malloc(self.len)
|
||
|
if not self.data:
|
||
|
raise MemoryError("unable to allocate array data.")
|
||
|
|
||
|
if self.dtype_is_object:
|
||
|
p = <PyObject **> self.data
|
||
|
for i in range(self.len / itemsize):
|
||
|
p[i] = Py_None
|
||
|
Py_INCREF(Py_None)
|
||
|
|
||
|
@cname('getbuffer')
|
||
|
def __getbuffer__(self, Py_buffer *info, int flags):
|
||
|
cdef int bufmode = -1
|
||
|
if self.mode == u"c":
|
||
|
bufmode = PyBUF_C_CONTIGUOUS | PyBUF_ANY_CONTIGUOUS
|
||
|
elif self.mode == u"fortran":
|
||
|
bufmode = PyBUF_F_CONTIGUOUS | PyBUF_ANY_CONTIGUOUS
|
||
|
if not (flags & bufmode):
|
||
|
raise ValueError("Can only create a buffer that is contiguous in memory.")
|
||
|
info.buf = self.data
|
||
|
info.len = self.len
|
||
|
info.ndim = self.ndim
|
||
|
info.shape = self._shape
|
||
|
info.strides = self._strides
|
||
|
info.suboffsets = NULL
|
||
|
info.itemsize = self.itemsize
|
||
|
info.readonly = 0
|
||
|
|
||
|
if flags & PyBUF_FORMAT:
|
||
|
info.format = self.format
|
||
|
else:
|
||
|
info.format = NULL
|
||
|
|
||
|
info.obj = self
|
||
|
|
||
|
__pyx_getbuffer = capsule(<void *> &__pyx_array_getbuffer, "getbuffer(obj, view, flags)")
|
||
|
|
||
|
def __dealloc__(array self):
|
||
|
if self.callback_free_data != NULL:
|
||
|
self.callback_free_data(self.data)
|
||
|
elif self.free_data:
|
||
|
if self.dtype_is_object:
|
||
|
refcount_objects_in_slice(self.data, self._shape,
|
||
|
self._strides, self.ndim, False)
|
||
|
free(self.data)
|
||
|
PyObject_Free(self._shape)
|
||
|
|
||
|
@property
|
||
|
def memview(self):
|
||
|
return self.get_memview()
|
||
|
|
||
|
@cname('get_memview')
|
||
|
cdef get_memview(self):
|
||
|
flags = PyBUF_ANY_CONTIGUOUS|PyBUF_FORMAT|PyBUF_WRITABLE
|
||
|
return memoryview(self, flags, self.dtype_is_object)
|
||
|
|
||
|
def __len__(self):
|
||
|
return self._shape[0]
|
||
|
|
||
|
def __getattr__(self, attr):
|
||
|
return getattr(self.memview, attr)
|
||
|
|
||
|
def __getitem__(self, item):
|
||
|
return self.memview[item]
|
||
|
|
||
|
def __setitem__(self, item, value):
|
||
|
self.memview[item] = value
|
||
|
|
||
|
|
||
|
@cname("__pyx_array_new")
|
||
|
cdef array array_cwrapper(tuple shape, Py_ssize_t itemsize, char *format,
|
||
|
char *mode, char *buf):
|
||
|
cdef array result
|
||
|
|
||
|
if buf == NULL:
|
||
|
result = array(shape, itemsize, format, mode.decode('ASCII'))
|
||
|
else:
|
||
|
result = array(shape, itemsize, format, mode.decode('ASCII'),
|
||
|
allocate_buffer=False)
|
||
|
result.data = buf
|
||
|
|
||
|
return result
|
||
|
|
||
|
|
||
|
#
|
||
|
### Memoryview constants and cython.view.memoryview class
|
||
|
#
|
||
|
|
||
|
# Disable generic_contiguous, as it makes trouble verifying contiguity:
|
||
|
# - 'contiguous' or '::1' means the dimension is contiguous with dtype
|
||
|
# - 'indirect_contiguous' means a contiguous list of pointers
|
||
|
# - dtype contiguous must be contiguous in the first or last dimension
|
||
|
# from the start, or from the dimension following the last indirect dimension
|
||
|
#
|
||
|
# e.g.
|
||
|
# int[::indirect_contiguous, ::contiguous, :]
|
||
|
#
|
||
|
# is valid (list of pointers to 2d fortran-contiguous array), but
|
||
|
#
|
||
|
# int[::generic_contiguous, ::contiguous, :]
|
||
|
#
|
||
|
# would mean you'd have assert dimension 0 to be indirect (and pointer contiguous) at runtime.
|
||
|
# So it doesn't bring any performance benefit, and it's only confusing.
|
||
|
|
||
|
@cname('__pyx_MemviewEnum')
|
||
|
cdef class Enum(object):
|
||
|
cdef object name
|
||
|
def __init__(self, name):
|
||
|
self.name = name
|
||
|
def __repr__(self):
|
||
|
return self.name
|
||
|
|
||
|
cdef generic = Enum("<strided and direct or indirect>")
|
||
|
cdef strided = Enum("<strided and direct>") # default
|
||
|
cdef indirect = Enum("<strided and indirect>")
|
||
|
# Disable generic_contiguous, as it is a troublemaker
|
||
|
#cdef generic_contiguous = Enum("<contiguous and direct or indirect>")
|
||
|
cdef contiguous = Enum("<contiguous and direct>")
|
||
|
cdef indirect_contiguous = Enum("<contiguous and indirect>")
|
||
|
|
||
|
# 'follow' is implied when the first or last axis is ::1
|
||
|
|
||
|
|
||
|
@cname('__pyx_align_pointer')
|
||
|
cdef void *align_pointer(void *memory, size_t alignment) nogil:
|
||
|
"Align pointer memory on a given boundary"
|
||
|
cdef Py_intptr_t aligned_p = <Py_intptr_t> memory
|
||
|
cdef size_t offset
|
||
|
|
||
|
with cython.cdivision(True):
|
||
|
offset = aligned_p % alignment
|
||
|
|
||
|
if offset > 0:
|
||
|
aligned_p += alignment - offset
|
||
|
|
||
|
return <void *> aligned_p
|
||
|
|
||
|
|
||
|
# pre-allocate thread locks for reuse
|
||
|
## note that this could be implemented in a more beautiful way in "normal" Cython,
|
||
|
## but this code gets merged into the user module and not everything works there.
|
||
|
DEF THREAD_LOCKS_PREALLOCATED = 8
|
||
|
cdef int __pyx_memoryview_thread_locks_used = 0
|
||
|
cdef PyThread_type_lock[THREAD_LOCKS_PREALLOCATED] __pyx_memoryview_thread_locks = [
|
||
|
PyThread_allocate_lock(),
|
||
|
PyThread_allocate_lock(),
|
||
|
PyThread_allocate_lock(),
|
||
|
PyThread_allocate_lock(),
|
||
|
PyThread_allocate_lock(),
|
||
|
PyThread_allocate_lock(),
|
||
|
PyThread_allocate_lock(),
|
||
|
PyThread_allocate_lock(),
|
||
|
]
|
||
|
|
||
|
|
||
|
@cname('__pyx_memoryview')
|
||
|
cdef class memoryview(object):
|
||
|
|
||
|
cdef object obj
|
||
|
cdef object _size
|
||
|
cdef object _array_interface
|
||
|
cdef PyThread_type_lock lock
|
||
|
# the following array will contain a single __pyx_atomic int with
|
||
|
# suitable alignment
|
||
|
cdef __pyx_atomic_int acquisition_count[2]
|
||
|
cdef __pyx_atomic_int *acquisition_count_aligned_p
|
||
|
cdef Py_buffer view
|
||
|
cdef int flags
|
||
|
cdef bint dtype_is_object
|
||
|
cdef __Pyx_TypeInfo *typeinfo
|
||
|
|
||
|
def __cinit__(memoryview self, object obj, int flags, bint dtype_is_object=False):
|
||
|
self.obj = obj
|
||
|
self.flags = flags
|
||
|
if type(self) is memoryview or obj is not None:
|
||
|
__Pyx_GetBuffer(obj, &self.view, flags)
|
||
|
if <PyObject *> self.view.obj == NULL:
|
||
|
(<__pyx_buffer *> &self.view).obj = Py_None
|
||
|
Py_INCREF(Py_None)
|
||
|
|
||
|
global __pyx_memoryview_thread_locks_used
|
||
|
if __pyx_memoryview_thread_locks_used < THREAD_LOCKS_PREALLOCATED:
|
||
|
self.lock = __pyx_memoryview_thread_locks[__pyx_memoryview_thread_locks_used]
|
||
|
__pyx_memoryview_thread_locks_used += 1
|
||
|
if self.lock is NULL:
|
||
|
self.lock = PyThread_allocate_lock()
|
||
|
if self.lock is NULL:
|
||
|
raise MemoryError
|
||
|
|
||
|
if flags & PyBUF_FORMAT:
|
||
|
self.dtype_is_object = (self.view.format[0] == b'O' and self.view.format[1] == b'\0')
|
||
|
else:
|
||
|
self.dtype_is_object = dtype_is_object
|
||
|
|
||
|
self.acquisition_count_aligned_p = <__pyx_atomic_int *> align_pointer(
|
||
|
<void *> &self.acquisition_count[0], sizeof(__pyx_atomic_int))
|
||
|
self.typeinfo = NULL
|
||
|
|
||
|
def __dealloc__(memoryview self):
|
||
|
if self.obj is not None:
|
||
|
__Pyx_ReleaseBuffer(&self.view)
|
||
|
elif (<__pyx_buffer *> &self.view).obj == Py_None:
|
||
|
# Undo the incref in __cinit__() above.
|
||
|
(<__pyx_buffer *> &self.view).obj = NULL
|
||
|
Py_DECREF(Py_None)
|
||
|
|
||
|
cdef int i
|
||
|
global __pyx_memoryview_thread_locks_used
|
||
|
if self.lock != NULL:
|
||
|
for i in range(__pyx_memoryview_thread_locks_used):
|
||
|
if __pyx_memoryview_thread_locks[i] is self.lock:
|
||
|
__pyx_memoryview_thread_locks_used -= 1
|
||
|
if i != __pyx_memoryview_thread_locks_used:
|
||
|
__pyx_memoryview_thread_locks[i], __pyx_memoryview_thread_locks[__pyx_memoryview_thread_locks_used] = (
|
||
|
__pyx_memoryview_thread_locks[__pyx_memoryview_thread_locks_used], __pyx_memoryview_thread_locks[i])
|
||
|
break
|
||
|
else:
|
||
|
PyThread_free_lock(self.lock)
|
||
|
|
||
|
cdef char *get_item_pointer(memoryview self, object index) except NULL:
|
||
|
cdef Py_ssize_t dim
|
||
|
cdef char *itemp = <char *> self.view.buf
|
||
|
|
||
|
for dim, idx in enumerate(index):
|
||
|
itemp = pybuffer_index(&self.view, itemp, idx, dim)
|
||
|
|
||
|
return itemp
|
||
|
|
||
|
#@cname('__pyx_memoryview_getitem')
|
||
|
def __getitem__(memoryview self, object index):
|
||
|
if index is Ellipsis:
|
||
|
return self
|
||
|
|
||
|
have_slices, indices = _unellipsify(index, self.view.ndim)
|
||
|
|
||
|
cdef char *itemp
|
||
|
if have_slices:
|
||
|
return memview_slice(self, indices)
|
||
|
else:
|
||
|
itemp = self.get_item_pointer(indices)
|
||
|
return self.convert_item_to_object(itemp)
|
||
|
|
||
|
def __setitem__(memoryview self, object index, object value):
|
||
|
if self.view.readonly:
|
||
|
raise TypeError("Cannot assign to read-only memoryview")
|
||
|
|
||
|
have_slices, index = _unellipsify(index, self.view.ndim)
|
||
|
|
||
|
if have_slices:
|
||
|
obj = self.is_slice(value)
|
||
|
if obj:
|
||
|
self.setitem_slice_assignment(self[index], obj)
|
||
|
else:
|
||
|
self.setitem_slice_assign_scalar(self[index], value)
|
||
|
else:
|
||
|
self.setitem_indexed(index, value)
|
||
|
|
||
|
cdef is_slice(self, obj):
|
||
|
if not isinstance(obj, memoryview):
|
||
|
try:
|
||
|
obj = memoryview(obj, self.flags & ~PyBUF_WRITABLE | PyBUF_ANY_CONTIGUOUS,
|
||
|
self.dtype_is_object)
|
||
|
except TypeError:
|
||
|
return None
|
||
|
|
||
|
return obj
|
||
|
|
||
|
cdef setitem_slice_assignment(self, dst, src):
|
||
|
cdef {{memviewslice_name}} dst_slice
|
||
|
cdef {{memviewslice_name}} src_slice
|
||
|
|
||
|
memoryview_copy_contents(get_slice_from_memview(src, &src_slice)[0],
|
||
|
get_slice_from_memview(dst, &dst_slice)[0],
|
||
|
src.ndim, dst.ndim, self.dtype_is_object)
|
||
|
|
||
|
cdef setitem_slice_assign_scalar(self, memoryview dst, value):
|
||
|
cdef int array[128]
|
||
|
cdef void *tmp = NULL
|
||
|
cdef void *item
|
||
|
|
||
|
cdef {{memviewslice_name}} *dst_slice
|
||
|
cdef {{memviewslice_name}} tmp_slice
|
||
|
dst_slice = get_slice_from_memview(dst, &tmp_slice)
|
||
|
|
||
|
if <size_t>self.view.itemsize > sizeof(array):
|
||
|
tmp = PyMem_Malloc(self.view.itemsize)
|
||
|
if tmp == NULL:
|
||
|
raise MemoryError
|
||
|
item = tmp
|
||
|
else:
|
||
|
item = <void *> array
|
||
|
|
||
|
try:
|
||
|
if self.dtype_is_object:
|
||
|
(<PyObject **> item)[0] = <PyObject *> value
|
||
|
else:
|
||
|
self.assign_item_from_object(<char *> item, value)
|
||
|
|
||
|
# It would be easy to support indirect dimensions, but it's easier
|
||
|
# to disallow :)
|
||
|
if self.view.suboffsets != NULL:
|
||
|
assert_direct_dimensions(self.view.suboffsets, self.view.ndim)
|
||
|
slice_assign_scalar(dst_slice, dst.view.ndim, self.view.itemsize,
|
||
|
item, self.dtype_is_object)
|
||
|
finally:
|
||
|
PyMem_Free(tmp)
|
||
|
|
||
|
cdef setitem_indexed(self, index, value):
|
||
|
cdef char *itemp = self.get_item_pointer(index)
|
||
|
self.assign_item_from_object(itemp, value)
|
||
|
|
||
|
cdef convert_item_to_object(self, char *itemp):
|
||
|
"""Only used if instantiated manually by the user, or if Cython doesn't
|
||
|
know how to convert the type"""
|
||
|
import struct
|
||
|
cdef bytes bytesitem
|
||
|
# Do a manual and complete check here instead of this easy hack
|
||
|
bytesitem = itemp[:self.view.itemsize]
|
||
|
try:
|
||
|
result = struct.unpack(self.view.format, bytesitem)
|
||
|
except struct.error:
|
||
|
raise ValueError("Unable to convert item to object")
|
||
|
else:
|
||
|
if len(self.view.format) == 1:
|
||
|
return result[0]
|
||
|
return result
|
||
|
|
||
|
cdef assign_item_from_object(self, char *itemp, object value):
|
||
|
"""Only used if instantiated manually by the user, or if Cython doesn't
|
||
|
know how to convert the type"""
|
||
|
import struct
|
||
|
cdef char c
|
||
|
cdef bytes bytesvalue
|
||
|
cdef Py_ssize_t i
|
||
|
|
||
|
if isinstance(value, tuple):
|
||
|
bytesvalue = struct.pack(self.view.format, *value)
|
||
|
else:
|
||
|
bytesvalue = struct.pack(self.view.format, value)
|
||
|
|
||
|
for i, c in enumerate(bytesvalue):
|
||
|
itemp[i] = c
|
||
|
|
||
|
@cname('getbuffer')
|
||
|
def __getbuffer__(self, Py_buffer *info, int flags):
|
||
|
if flags & PyBUF_WRITABLE and self.view.readonly:
|
||
|
raise ValueError("Cannot create writable memory view from read-only memoryview")
|
||
|
|
||
|
if flags & PyBUF_ND:
|
||
|
info.shape = self.view.shape
|
||
|
else:
|
||
|
info.shape = NULL
|
||
|
|
||
|
if flags & PyBUF_STRIDES:
|
||
|
info.strides = self.view.strides
|
||
|
else:
|
||
|
info.strides = NULL
|
||
|
|
||
|
if flags & PyBUF_INDIRECT:
|
||
|
info.suboffsets = self.view.suboffsets
|
||
|
else:
|
||
|
info.suboffsets = NULL
|
||
|
|
||
|
if flags & PyBUF_FORMAT:
|
||
|
info.format = self.view.format
|
||
|
else:
|
||
|
info.format = NULL
|
||
|
|
||
|
info.buf = self.view.buf
|
||
|
info.ndim = self.view.ndim
|
||
|
info.itemsize = self.view.itemsize
|
||
|
info.len = self.view.len
|
||
|
info.readonly = self.view.readonly
|
||
|
info.obj = self
|
||
|
|
||
|
__pyx_getbuffer = capsule(<void *> &__pyx_memoryview_getbuffer, "getbuffer(obj, view, flags)")
|
||
|
|
||
|
# Some properties that have the same semantics as in NumPy
|
||
|
@property
|
||
|
def T(self):
|
||
|
cdef _memoryviewslice result = memoryview_copy(self)
|
||
|
transpose_memslice(&result.from_slice)
|
||
|
return result
|
||
|
|
||
|
@property
|
||
|
def base(self):
|
||
|
return self.obj
|
||
|
|
||
|
@property
|
||
|
def shape(self):
|
||
|
return tuple([length for length in self.view.shape[:self.view.ndim]])
|
||
|
|
||
|
@property
|
||
|
def strides(self):
|
||
|
if self.view.strides == NULL:
|
||
|
# Note: we always ask for strides, so if this is not set it's a bug
|
||
|
raise ValueError("Buffer view does not expose strides")
|
||
|
|
||
|
return tuple([stride for stride in self.view.strides[:self.view.ndim]])
|
||
|
|
||
|
@property
|
||
|
def suboffsets(self):
|
||
|
if self.view.suboffsets == NULL:
|
||
|
return (-1,) * self.view.ndim
|
||
|
|
||
|
return tuple([suboffset for suboffset in self.view.suboffsets[:self.view.ndim]])
|
||
|
|
||
|
@property
|
||
|
def ndim(self):
|
||
|
return self.view.ndim
|
||
|
|
||
|
@property
|
||
|
def itemsize(self):
|
||
|
return self.view.itemsize
|
||
|
|
||
|
@property
|
||
|
def nbytes(self):
|
||
|
return self.size * self.view.itemsize
|
||
|
|
||
|
@property
|
||
|
def size(self):
|
||
|
if self._size is None:
|
||
|
result = 1
|
||
|
|
||
|
for length in self.view.shape[:self.view.ndim]:
|
||
|
result *= length
|
||
|
|
||
|
self._size = result
|
||
|
|
||
|
return self._size
|
||
|
|
||
|
def __len__(self):
|
||
|
if self.view.ndim >= 1:
|
||
|
return self.view.shape[0]
|
||
|
|
||
|
return 0
|
||
|
|
||
|
def __repr__(self):
|
||
|
return "<MemoryView of %r at 0x%x>" % (self.base.__class__.__name__,
|
||
|
id(self))
|
||
|
|
||
|
def __str__(self):
|
||
|
return "<MemoryView of %r object>" % (self.base.__class__.__name__,)
|
||
|
|
||
|
# Support the same attributes as memoryview slices
|
||
|
def is_c_contig(self):
|
||
|
cdef {{memviewslice_name}} *mslice
|
||
|
cdef {{memviewslice_name}} tmp
|
||
|
mslice = get_slice_from_memview(self, &tmp)
|
||
|
return slice_is_contig(mslice[0], 'C', self.view.ndim)
|
||
|
|
||
|
def is_f_contig(self):
|
||
|
cdef {{memviewslice_name}} *mslice
|
||
|
cdef {{memviewslice_name}} tmp
|
||
|
mslice = get_slice_from_memview(self, &tmp)
|
||
|
return slice_is_contig(mslice[0], 'F', self.view.ndim)
|
||
|
|
||
|
def copy(self):
|
||
|
cdef {{memviewslice_name}} mslice
|
||
|
cdef int flags = self.flags & ~PyBUF_F_CONTIGUOUS
|
||
|
|
||
|
slice_copy(self, &mslice)
|
||
|
mslice = slice_copy_contig(&mslice, "c", self.view.ndim,
|
||
|
self.view.itemsize,
|
||
|
flags|PyBUF_C_CONTIGUOUS,
|
||
|
self.dtype_is_object)
|
||
|
|
||
|
return memoryview_copy_from_slice(self, &mslice)
|
||
|
|
||
|
def copy_fortran(self):
|
||
|
cdef {{memviewslice_name}} src, dst
|
||
|
cdef int flags = self.flags & ~PyBUF_C_CONTIGUOUS
|
||
|
|
||
|
slice_copy(self, &src)
|
||
|
dst = slice_copy_contig(&src, "fortran", self.view.ndim,
|
||
|
self.view.itemsize,
|
||
|
flags|PyBUF_F_CONTIGUOUS,
|
||
|
self.dtype_is_object)
|
||
|
|
||
|
return memoryview_copy_from_slice(self, &dst)
|
||
|
|
||
|
|
||
|
@cname('__pyx_memoryview_new')
|
||
|
cdef memoryview_cwrapper(object o, int flags, bint dtype_is_object, __Pyx_TypeInfo *typeinfo):
|
||
|
cdef memoryview result = memoryview(o, flags, dtype_is_object)
|
||
|
result.typeinfo = typeinfo
|
||
|
return result
|
||
|
|
||
|
@cname('__pyx_memoryview_check')
|
||
|
cdef inline bint memoryview_check(object o):
|
||
|
return isinstance(o, memoryview)
|
||
|
|
||
|
cdef tuple _unellipsify(object index, int ndim):
|
||
|
"""
|
||
|
Replace all ellipses with full slices and fill incomplete indices with
|
||
|
full slices.
|
||
|
"""
|
||
|
if not isinstance(index, tuple):
|
||
|
tup = (index,)
|
||
|
else:
|
||
|
tup = index
|
||
|
|
||
|
result = []
|
||
|
have_slices = False
|
||
|
seen_ellipsis = False
|
||
|
for idx, item in enumerate(tup):
|
||
|
if item is Ellipsis:
|
||
|
if not seen_ellipsis:
|
||
|
result.extend([slice(None)] * (ndim - len(tup) + 1))
|
||
|
seen_ellipsis = True
|
||
|
else:
|
||
|
result.append(slice(None))
|
||
|
have_slices = True
|
||
|
else:
|
||
|
if not isinstance(item, slice) and not PyIndex_Check(item):
|
||
|
raise TypeError("Cannot index with type '%s'" % type(item))
|
||
|
|
||
|
have_slices = have_slices or isinstance(item, slice)
|
||
|
result.append(item)
|
||
|
|
||
|
nslices = ndim - len(result)
|
||
|
if nslices:
|
||
|
result.extend([slice(None)] * nslices)
|
||
|
|
||
|
return have_slices or nslices, tuple(result)
|
||
|
|
||
|
cdef assert_direct_dimensions(Py_ssize_t *suboffsets, int ndim):
|
||
|
for suboffset in suboffsets[:ndim]:
|
||
|
if suboffset >= 0:
|
||
|
raise ValueError("Indirect dimensions not supported")
|
||
|
|
||
|
#
|
||
|
### Slicing a memoryview
|
||
|
#
|
||
|
|
||
|
@cname('__pyx_memview_slice')
|
||
|
cdef memoryview memview_slice(memoryview memview, object indices):
|
||
|
cdef int new_ndim = 0, suboffset_dim = -1, dim
|
||
|
cdef bint negative_step
|
||
|
cdef {{memviewslice_name}} src, dst
|
||
|
cdef {{memviewslice_name}} *p_src
|
||
|
|
||
|
# dst is copied by value in memoryview_fromslice -- initialize it
|
||
|
# src is never copied
|
||
|
memset(&dst, 0, sizeof(dst))
|
||
|
|
||
|
cdef _memoryviewslice memviewsliceobj
|
||
|
|
||
|
assert memview.view.ndim > 0
|
||
|
|
||
|
if isinstance(memview, _memoryviewslice):
|
||
|
memviewsliceobj = memview
|
||
|
p_src = &memviewsliceobj.from_slice
|
||
|
else:
|
||
|
slice_copy(memview, &src)
|
||
|
p_src = &src
|
||
|
|
||
|
# Note: don't use variable src at this point
|
||
|
# SubNote: we should be able to declare variables in blocks...
|
||
|
|
||
|
# memoryview_fromslice() will inc our dst slice
|
||
|
dst.memview = p_src.memview
|
||
|
dst.data = p_src.data
|
||
|
|
||
|
# Put everything in temps to avoid this bloody warning:
|
||
|
# "Argument evaluation order in C function call is undefined and
|
||
|
# may not be as expected"
|
||
|
cdef {{memviewslice_name}} *p_dst = &dst
|
||
|
cdef int *p_suboffset_dim = &suboffset_dim
|
||
|
cdef Py_ssize_t start, stop, step
|
||
|
cdef bint have_start, have_stop, have_step
|
||
|
|
||
|
for dim, index in enumerate(indices):
|
||
|
if PyIndex_Check(index):
|
||
|
slice_memviewslice(
|
||
|
p_dst, p_src.shape[dim], p_src.strides[dim], p_src.suboffsets[dim],
|
||
|
dim, new_ndim, p_suboffset_dim,
|
||
|
index, 0, 0, # start, stop, step
|
||
|
0, 0, 0, # have_{start,stop,step}
|
||
|
False)
|
||
|
elif index is None:
|
||
|
p_dst.shape[new_ndim] = 1
|
||
|
p_dst.strides[new_ndim] = 0
|
||
|
p_dst.suboffsets[new_ndim] = -1
|
||
|
new_ndim += 1
|
||
|
else:
|
||
|
start = index.start or 0
|
||
|
stop = index.stop or 0
|
||
|
step = index.step or 0
|
||
|
|
||
|
have_start = index.start is not None
|
||
|
have_stop = index.stop is not None
|
||
|
have_step = index.step is not None
|
||
|
|
||
|
slice_memviewslice(
|
||
|
p_dst, p_src.shape[dim], p_src.strides[dim], p_src.suboffsets[dim],
|
||
|
dim, new_ndim, p_suboffset_dim,
|
||
|
start, stop, step,
|
||
|
have_start, have_stop, have_step,
|
||
|
True)
|
||
|
new_ndim += 1
|
||
|
|
||
|
if isinstance(memview, _memoryviewslice):
|
||
|
return memoryview_fromslice(dst, new_ndim,
|
||
|
memviewsliceobj.to_object_func,
|
||
|
memviewsliceobj.to_dtype_func,
|
||
|
memview.dtype_is_object)
|
||
|
else:
|
||
|
return memoryview_fromslice(dst, new_ndim, NULL, NULL,
|
||
|
memview.dtype_is_object)
|
||
|
|
||
|
|
||
|
#
|
||
|
### Slicing in a single dimension of a memoryviewslice
|
||
|
#
|
||
|
|
||
|
cdef extern from "<stdlib.h>":
|
||
|
void abort() nogil
|
||
|
void printf(char *s, ...) nogil
|
||
|
|
||
|
cdef extern from "<stdio.h>":
|
||
|
ctypedef struct FILE
|
||
|
FILE *stderr
|
||
|
int fputs(char *s, FILE *stream)
|
||
|
|
||
|
cdef extern from "pystate.h":
|
||
|
void PyThreadState_Get() nogil
|
||
|
|
||
|
# These are not actually nogil, but we check for the GIL before calling them
|
||
|
void PyErr_SetString(PyObject *type, char *msg) nogil
|
||
|
PyObject *PyErr_Format(PyObject *exc, char *msg, ...) nogil
|
||
|
|
||
|
@cname('__pyx_memoryview_slice_memviewslice')
|
||
|
cdef int slice_memviewslice(
|
||
|
{{memviewslice_name}} *dst,
|
||
|
Py_ssize_t shape, Py_ssize_t stride, Py_ssize_t suboffset,
|
||
|
int dim, int new_ndim, int *suboffset_dim,
|
||
|
Py_ssize_t start, Py_ssize_t stop, Py_ssize_t step,
|
||
|
int have_start, int have_stop, int have_step,
|
||
|
bint is_slice) nogil except -1:
|
||
|
"""
|
||
|
Create a new slice dst given slice src.
|
||
|
|
||
|
dim - the current src dimension (indexing will make dimensions
|
||
|
disappear)
|
||
|
new_dim - the new dst dimension
|
||
|
suboffset_dim - pointer to a single int initialized to -1 to keep track of
|
||
|
where slicing offsets should be added
|
||
|
"""
|
||
|
|
||
|
cdef Py_ssize_t new_shape
|
||
|
cdef bint negative_step
|
||
|
|
||
|
if not is_slice:
|
||
|
# index is a normal integer-like index
|
||
|
if start < 0:
|
||
|
start += shape
|
||
|
if not 0 <= start < shape:
|
||
|
_err_dim(IndexError, "Index out of bounds (axis %d)", dim)
|
||
|
else:
|
||
|
# index is a slice
|
||
|
negative_step = have_step != 0 and step < 0
|
||
|
|
||
|
if have_step and step == 0:
|
||
|
_err_dim(ValueError, "Step may not be zero (axis %d)", dim)
|
||
|
|
||
|
# check our bounds and set defaults
|
||
|
if have_start:
|
||
|
if start < 0:
|
||
|
start += shape
|
||
|
if start < 0:
|
||
|
start = 0
|
||
|
elif start >= shape:
|
||
|
if negative_step:
|
||
|
start = shape - 1
|
||
|
else:
|
||
|
start = shape
|
||
|
else:
|
||
|
if negative_step:
|
||
|
start = shape - 1
|
||
|
else:
|
||
|
start = 0
|
||
|
|
||
|
if have_stop:
|
||
|
if stop < 0:
|
||
|
stop += shape
|
||
|
if stop < 0:
|
||
|
stop = 0
|
||
|
elif stop > shape:
|
||
|
stop = shape
|
||
|
else:
|
||
|
if negative_step:
|
||
|
stop = -1
|
||
|
else:
|
||
|
stop = shape
|
||
|
|
||
|
if not have_step:
|
||
|
step = 1
|
||
|
|
||
|
# len = ceil( (stop - start) / step )
|
||
|
with cython.cdivision(True):
|
||
|
new_shape = (stop - start) // step
|
||
|
|
||
|
if (stop - start) - step * new_shape:
|
||
|
new_shape += 1
|
||
|
|
||
|
if new_shape < 0:
|
||
|
new_shape = 0
|
||
|
|
||
|
# shape/strides/suboffsets
|
||
|
dst.strides[new_ndim] = stride * step
|
||
|
dst.shape[new_ndim] = new_shape
|
||
|
dst.suboffsets[new_ndim] = suboffset
|
||
|
|
||
|
# Add the slicing or idexing offsets to the right suboffset or base data *
|
||
|
if suboffset_dim[0] < 0:
|
||
|
dst.data += start * stride
|
||
|
else:
|
||
|
dst.suboffsets[suboffset_dim[0]] += start * stride
|
||
|
|
||
|
if suboffset >= 0:
|
||
|
if not is_slice:
|
||
|
if new_ndim == 0:
|
||
|
dst.data = (<char **> dst.data)[0] + suboffset
|
||
|
else:
|
||
|
_err_dim(IndexError, "All dimensions preceding dimension %d "
|
||
|
"must be indexed and not sliced", dim)
|
||
|
else:
|
||
|
suboffset_dim[0] = new_ndim
|
||
|
|
||
|
return 0
|
||
|
|
||
|
#
|
||
|
### Index a memoryview
|
||
|
#
|
||
|
@cname('__pyx_pybuffer_index')
|
||
|
cdef char *pybuffer_index(Py_buffer *view, char *bufp, Py_ssize_t index,
|
||
|
Py_ssize_t dim) except NULL:
|
||
|
cdef Py_ssize_t shape, stride, suboffset = -1
|
||
|
cdef Py_ssize_t itemsize = view.itemsize
|
||
|
cdef char *resultp
|
||
|
|
||
|
if view.ndim == 0:
|
||
|
shape = view.len / itemsize
|
||
|
stride = itemsize
|
||
|
else:
|
||
|
shape = view.shape[dim]
|
||
|
stride = view.strides[dim]
|
||
|
if view.suboffsets != NULL:
|
||
|
suboffset = view.suboffsets[dim]
|
||
|
|
||
|
if index < 0:
|
||
|
index += view.shape[dim]
|
||
|
if index < 0:
|
||
|
raise IndexError("Out of bounds on buffer access (axis %d)" % dim)
|
||
|
|
||
|
if index >= shape:
|
||
|
raise IndexError("Out of bounds on buffer access (axis %d)" % dim)
|
||
|
|
||
|
resultp = bufp + index * stride
|
||
|
if suboffset >= 0:
|
||
|
resultp = (<char **> resultp)[0] + suboffset
|
||
|
|
||
|
return resultp
|
||
|
|
||
|
#
|
||
|
### Transposing a memoryviewslice
|
||
|
#
|
||
|
@cname('__pyx_memslice_transpose')
|
||
|
cdef int transpose_memslice({{memviewslice_name}} *memslice) nogil except 0:
|
||
|
cdef int ndim = memslice.memview.view.ndim
|
||
|
|
||
|
cdef Py_ssize_t *shape = memslice.shape
|
||
|
cdef Py_ssize_t *strides = memslice.strides
|
||
|
|
||
|
# reverse strides and shape
|
||
|
cdef int i, j
|
||
|
for i in range(ndim / 2):
|
||
|
j = ndim - 1 - i
|
||
|
strides[i], strides[j] = strides[j], strides[i]
|
||
|
shape[i], shape[j] = shape[j], shape[i]
|
||
|
|
||
|
if memslice.suboffsets[i] >= 0 or memslice.suboffsets[j] >= 0:
|
||
|
_err(ValueError, "Cannot transpose memoryview with indirect dimensions")
|
||
|
|
||
|
return 1
|
||
|
|
||
|
#
|
||
|
### Creating new memoryview objects from slices and memoryviews
|
||
|
#
|
||
|
@cname('__pyx_memoryviewslice')
|
||
|
cdef class _memoryviewslice(memoryview):
|
||
|
"Internal class for passing memoryview slices to Python"
|
||
|
|
||
|
# We need this to keep our shape/strides/suboffset pointers valid
|
||
|
cdef {{memviewslice_name}} from_slice
|
||
|
# We need this only to print it's class' name
|
||
|
cdef object from_object
|
||
|
|
||
|
cdef object (*to_object_func)(char *)
|
||
|
cdef int (*to_dtype_func)(char *, object) except 0
|
||
|
|
||
|
def __dealloc__(self):
|
||
|
__PYX_XDEC_MEMVIEW(&self.from_slice, 1)
|
||
|
|
||
|
cdef convert_item_to_object(self, char *itemp):
|
||
|
if self.to_object_func != NULL:
|
||
|
return self.to_object_func(itemp)
|
||
|
else:
|
||
|
return memoryview.convert_item_to_object(self, itemp)
|
||
|
|
||
|
cdef assign_item_from_object(self, char *itemp, object value):
|
||
|
if self.to_dtype_func != NULL:
|
||
|
self.to_dtype_func(itemp, value)
|
||
|
else:
|
||
|
memoryview.assign_item_from_object(self, itemp, value)
|
||
|
|
||
|
@property
|
||
|
def base(self):
|
||
|
return self.from_object
|
||
|
|
||
|
__pyx_getbuffer = capsule(<void *> &__pyx_memoryview_getbuffer, "getbuffer(obj, view, flags)")
|
||
|
|
||
|
|
||
|
@cname('__pyx_memoryview_fromslice')
|
||
|
cdef memoryview_fromslice({{memviewslice_name}} memviewslice,
|
||
|
int ndim,
|
||
|
object (*to_object_func)(char *),
|
||
|
int (*to_dtype_func)(char *, object) except 0,
|
||
|
bint dtype_is_object):
|
||
|
|
||
|
cdef _memoryviewslice result
|
||
|
|
||
|
if <PyObject *> memviewslice.memview == Py_None:
|
||
|
return None
|
||
|
|
||
|
# assert 0 < ndim <= memviewslice.memview.view.ndim, (
|
||
|
# ndim, memviewslice.memview.view.ndim)
|
||
|
|
||
|
result = _memoryviewslice(None, 0, dtype_is_object)
|
||
|
|
||
|
result.from_slice = memviewslice
|
||
|
__PYX_INC_MEMVIEW(&memviewslice, 1)
|
||
|
|
||
|
result.from_object = (<memoryview> memviewslice.memview).base
|
||
|
result.typeinfo = memviewslice.memview.typeinfo
|
||
|
|
||
|
result.view = memviewslice.memview.view
|
||
|
result.view.buf = <void *> memviewslice.data
|
||
|
result.view.ndim = ndim
|
||
|
(<__pyx_buffer *> &result.view).obj = Py_None
|
||
|
Py_INCREF(Py_None)
|
||
|
|
||
|
if (<memoryview>memviewslice.memview).flags & PyBUF_WRITABLE:
|
||
|
result.flags = PyBUF_RECORDS
|
||
|
else:
|
||
|
result.flags = PyBUF_RECORDS_RO
|
||
|
|
||
|
result.view.shape = <Py_ssize_t *> result.from_slice.shape
|
||
|
result.view.strides = <Py_ssize_t *> result.from_slice.strides
|
||
|
|
||
|
# only set suboffsets if actually used, otherwise set to NULL to improve compatibility
|
||
|
result.view.suboffsets = NULL
|
||
|
for suboffset in result.from_slice.suboffsets[:ndim]:
|
||
|
if suboffset >= 0:
|
||
|
result.view.suboffsets = <Py_ssize_t *> result.from_slice.suboffsets
|
||
|
break
|
||
|
|
||
|
result.view.len = result.view.itemsize
|
||
|
for length in result.view.shape[:ndim]:
|
||
|
result.view.len *= length
|
||
|
|
||
|
result.to_object_func = to_object_func
|
||
|
result.to_dtype_func = to_dtype_func
|
||
|
|
||
|
return result
|
||
|
|
||
|
@cname('__pyx_memoryview_get_slice_from_memoryview')
|
||
|
cdef {{memviewslice_name}} *get_slice_from_memview(memoryview memview,
|
||
|
{{memviewslice_name}} *mslice) except NULL:
|
||
|
cdef _memoryviewslice obj
|
||
|
if isinstance(memview, _memoryviewslice):
|
||
|
obj = memview
|
||
|
return &obj.from_slice
|
||
|
else:
|
||
|
slice_copy(memview, mslice)
|
||
|
return mslice
|
||
|
|
||
|
@cname('__pyx_memoryview_slice_copy')
|
||
|
cdef void slice_copy(memoryview memview, {{memviewslice_name}} *dst):
|
||
|
cdef int dim
|
||
|
cdef (Py_ssize_t*) shape, strides, suboffsets
|
||
|
|
||
|
shape = memview.view.shape
|
||
|
strides = memview.view.strides
|
||
|
suboffsets = memview.view.suboffsets
|
||
|
|
||
|
dst.memview = <__pyx_memoryview *> memview
|
||
|
dst.data = <char *> memview.view.buf
|
||
|
|
||
|
for dim in range(memview.view.ndim):
|
||
|
dst.shape[dim] = shape[dim]
|
||
|
dst.strides[dim] = strides[dim]
|
||
|
dst.suboffsets[dim] = suboffsets[dim] if suboffsets else -1
|
||
|
|
||
|
@cname('__pyx_memoryview_copy_object')
|
||
|
cdef memoryview_copy(memoryview memview):
|
||
|
"Create a new memoryview object"
|
||
|
cdef {{memviewslice_name}} memviewslice
|
||
|
slice_copy(memview, &memviewslice)
|
||
|
return memoryview_copy_from_slice(memview, &memviewslice)
|
||
|
|
||
|
@cname('__pyx_memoryview_copy_object_from_slice')
|
||
|
cdef memoryview_copy_from_slice(memoryview memview, {{memviewslice_name}} *memviewslice):
|
||
|
"""
|
||
|
Create a new memoryview object from a given memoryview object and slice.
|
||
|
"""
|
||
|
cdef object (*to_object_func)(char *)
|
||
|
cdef int (*to_dtype_func)(char *, object) except 0
|
||
|
|
||
|
if isinstance(memview, _memoryviewslice):
|
||
|
to_object_func = (<_memoryviewslice> memview).to_object_func
|
||
|
to_dtype_func = (<_memoryviewslice> memview).to_dtype_func
|
||
|
else:
|
||
|
to_object_func = NULL
|
||
|
to_dtype_func = NULL
|
||
|
|
||
|
return memoryview_fromslice(memviewslice[0], memview.view.ndim,
|
||
|
to_object_func, to_dtype_func,
|
||
|
memview.dtype_is_object)
|
||
|
|
||
|
|
||
|
#
|
||
|
### Copy the contents of a memoryview slices
|
||
|
#
|
||
|
cdef Py_ssize_t abs_py_ssize_t(Py_ssize_t arg) nogil:
|
||
|
if arg < 0:
|
||
|
return -arg
|
||
|
else:
|
||
|
return arg
|
||
|
|
||
|
@cname('__pyx_get_best_slice_order')
|
||
|
cdef char get_best_order({{memviewslice_name}} *mslice, int ndim) nogil:
|
||
|
"""
|
||
|
Figure out the best memory access order for a given slice.
|
||
|
"""
|
||
|
cdef int i
|
||
|
cdef Py_ssize_t c_stride = 0
|
||
|
cdef Py_ssize_t f_stride = 0
|
||
|
|
||
|
for i in range(ndim - 1, -1, -1):
|
||
|
if mslice.shape[i] > 1:
|
||
|
c_stride = mslice.strides[i]
|
||
|
break
|
||
|
|
||
|
for i in range(ndim):
|
||
|
if mslice.shape[i] > 1:
|
||
|
f_stride = mslice.strides[i]
|
||
|
break
|
||
|
|
||
|
if abs_py_ssize_t(c_stride) <= abs_py_ssize_t(f_stride):
|
||
|
return 'C'
|
||
|
else:
|
||
|
return 'F'
|
||
|
|
||
|
@cython.cdivision(True)
|
||
|
cdef void _copy_strided_to_strided(char *src_data, Py_ssize_t *src_strides,
|
||
|
char *dst_data, Py_ssize_t *dst_strides,
|
||
|
Py_ssize_t *src_shape, Py_ssize_t *dst_shape,
|
||
|
int ndim, size_t itemsize) nogil:
|
||
|
# Note: src_extent is 1 if we're broadcasting
|
||
|
# dst_extent always >= src_extent as we don't do reductions
|
||
|
cdef Py_ssize_t i
|
||
|
cdef Py_ssize_t src_extent = src_shape[0]
|
||
|
cdef Py_ssize_t dst_extent = dst_shape[0]
|
||
|
cdef Py_ssize_t src_stride = src_strides[0]
|
||
|
cdef Py_ssize_t dst_stride = dst_strides[0]
|
||
|
|
||
|
if ndim == 1:
|
||
|
if (src_stride > 0 and dst_stride > 0 and
|
||
|
<size_t> src_stride == itemsize == <size_t> dst_stride):
|
||
|
memcpy(dst_data, src_data, itemsize * dst_extent)
|
||
|
else:
|
||
|
for i in range(dst_extent):
|
||
|
memcpy(dst_data, src_data, itemsize)
|
||
|
src_data += src_stride
|
||
|
dst_data += dst_stride
|
||
|
else:
|
||
|
for i in range(dst_extent):
|
||
|
_copy_strided_to_strided(src_data, src_strides + 1,
|
||
|
dst_data, dst_strides + 1,
|
||
|
src_shape + 1, dst_shape + 1,
|
||
|
ndim - 1, itemsize)
|
||
|
src_data += src_stride
|
||
|
dst_data += dst_stride
|
||
|
|
||
|
cdef void copy_strided_to_strided({{memviewslice_name}} *src,
|
||
|
{{memviewslice_name}} *dst,
|
||
|
int ndim, size_t itemsize) nogil:
|
||
|
_copy_strided_to_strided(src.data, src.strides, dst.data, dst.strides,
|
||
|
src.shape, dst.shape, ndim, itemsize)
|
||
|
|
||
|
@cname('__pyx_memoryview_slice_get_size')
|
||
|
cdef Py_ssize_t slice_get_size({{memviewslice_name}} *src, int ndim) nogil:
|
||
|
"Return the size of the memory occupied by the slice in number of bytes"
|
||
|
cdef Py_ssize_t shape, size = src.memview.view.itemsize
|
||
|
|
||
|
for shape in src.shape[:ndim]:
|
||
|
size *= shape
|
||
|
|
||
|
return size
|
||
|
|
||
|
@cname('__pyx_fill_contig_strides_array')
|
||
|
cdef Py_ssize_t fill_contig_strides_array(
|
||
|
Py_ssize_t *shape, Py_ssize_t *strides, Py_ssize_t stride,
|
||
|
int ndim, char order) nogil:
|
||
|
"""
|
||
|
Fill the strides array for a slice with C or F contiguous strides.
|
||
|
This is like PyBuffer_FillContiguousStrides, but compatible with py < 2.6
|
||
|
"""
|
||
|
cdef int idx
|
||
|
|
||
|
if order == 'F':
|
||
|
for idx in range(ndim):
|
||
|
strides[idx] = stride
|
||
|
stride *= shape[idx]
|
||
|
else:
|
||
|
for idx in range(ndim - 1, -1, -1):
|
||
|
strides[idx] = stride
|
||
|
stride *= shape[idx]
|
||
|
|
||
|
return stride
|
||
|
|
||
|
@cname('__pyx_memoryview_copy_data_to_temp')
|
||
|
cdef void *copy_data_to_temp({{memviewslice_name}} *src,
|
||
|
{{memviewslice_name}} *tmpslice,
|
||
|
char order,
|
||
|
int ndim) nogil except NULL:
|
||
|
"""
|
||
|
Copy a direct slice to temporary contiguous memory. The caller should free
|
||
|
the result when done.
|
||
|
"""
|
||
|
cdef int i
|
||
|
cdef void *result
|
||
|
|
||
|
cdef size_t itemsize = src.memview.view.itemsize
|
||
|
cdef size_t size = slice_get_size(src, ndim)
|
||
|
|
||
|
result = malloc(size)
|
||
|
if not result:
|
||
|
_err(MemoryError, NULL)
|
||
|
|
||
|
# tmpslice[0] = src
|
||
|
tmpslice.data = <char *> result
|
||
|
tmpslice.memview = src.memview
|
||
|
for i in range(ndim):
|
||
|
tmpslice.shape[i] = src.shape[i]
|
||
|
tmpslice.suboffsets[i] = -1
|
||
|
|
||
|
fill_contig_strides_array(&tmpslice.shape[0], &tmpslice.strides[0], itemsize,
|
||
|
ndim, order)
|
||
|
|
||
|
# We need to broadcast strides again
|
||
|
for i in range(ndim):
|
||
|
if tmpslice.shape[i] == 1:
|
||
|
tmpslice.strides[i] = 0
|
||
|
|
||
|
if slice_is_contig(src[0], order, ndim):
|
||
|
memcpy(result, src.data, size)
|
||
|
else:
|
||
|
copy_strided_to_strided(src, tmpslice, ndim, itemsize)
|
||
|
|
||
|
return result
|
||
|
|
||
|
# Use 'with gil' functions and avoid 'with gil' blocks, as the code within the blocks
|
||
|
# has temporaries that need the GIL to clean up
|
||
|
@cname('__pyx_memoryview_err_extents')
|
||
|
cdef int _err_extents(int i, Py_ssize_t extent1,
|
||
|
Py_ssize_t extent2) except -1 with gil:
|
||
|
raise ValueError("got differing extents in dimension %d (got %d and %d)" %
|
||
|
(i, extent1, extent2))
|
||
|
|
||
|
@cname('__pyx_memoryview_err_dim')
|
||
|
cdef int _err_dim(object error, char *msg, int dim) except -1 with gil:
|
||
|
raise error(msg.decode('ascii') % dim)
|
||
|
|
||
|
@cname('__pyx_memoryview_err')
|
||
|
cdef int _err(object error, char *msg) except -1 with gil:
|
||
|
if msg != NULL:
|
||
|
raise error(msg.decode('ascii'))
|
||
|
else:
|
||
|
raise error
|
||
|
|
||
|
@cname('__pyx_memoryview_copy_contents')
|
||
|
cdef int memoryview_copy_contents({{memviewslice_name}} src,
|
||
|
{{memviewslice_name}} dst,
|
||
|
int src_ndim, int dst_ndim,
|
||
|
bint dtype_is_object) nogil except -1:
|
||
|
"""
|
||
|
Copy memory from slice src to slice dst.
|
||
|
Check for overlapping memory and verify the shapes.
|
||
|
"""
|
||
|
cdef void *tmpdata = NULL
|
||
|
cdef size_t itemsize = src.memview.view.itemsize
|
||
|
cdef int i
|
||
|
cdef char order = get_best_order(&src, src_ndim)
|
||
|
cdef bint broadcasting = False
|
||
|
cdef bint direct_copy = False
|
||
|
cdef {{memviewslice_name}} tmp
|
||
|
|
||
|
if src_ndim < dst_ndim:
|
||
|
broadcast_leading(&src, src_ndim, dst_ndim)
|
||
|
elif dst_ndim < src_ndim:
|
||
|
broadcast_leading(&dst, dst_ndim, src_ndim)
|
||
|
|
||
|
cdef int ndim = max(src_ndim, dst_ndim)
|
||
|
|
||
|
for i in range(ndim):
|
||
|
if src.shape[i] != dst.shape[i]:
|
||
|
if src.shape[i] == 1:
|
||
|
broadcasting = True
|
||
|
src.strides[i] = 0
|
||
|
else:
|
||
|
_err_extents(i, dst.shape[i], src.shape[i])
|
||
|
|
||
|
if src.suboffsets[i] >= 0:
|
||
|
_err_dim(ValueError, "Dimension %d is not direct", i)
|
||
|
|
||
|
if slices_overlap(&src, &dst, ndim, itemsize):
|
||
|
# slices overlap, copy to temp, copy temp to dst
|
||
|
if not slice_is_contig(src, order, ndim):
|
||
|
order = get_best_order(&dst, ndim)
|
||
|
|
||
|
tmpdata = copy_data_to_temp(&src, &tmp, order, ndim)
|
||
|
src = tmp
|
||
|
|
||
|
if not broadcasting:
|
||
|
# See if both slices have equal contiguity, in that case perform a
|
||
|
# direct copy. This only works when we are not broadcasting.
|
||
|
if slice_is_contig(src, 'C', ndim):
|
||
|
direct_copy = slice_is_contig(dst, 'C', ndim)
|
||
|
elif slice_is_contig(src, 'F', ndim):
|
||
|
direct_copy = slice_is_contig(dst, 'F', ndim)
|
||
|
|
||
|
if direct_copy:
|
||
|
# Contiguous slices with same order
|
||
|
refcount_copying(&dst, dtype_is_object, ndim, False)
|
||
|
memcpy(dst.data, src.data, slice_get_size(&src, ndim))
|
||
|
refcount_copying(&dst, dtype_is_object, ndim, True)
|
||
|
free(tmpdata)
|
||
|
return 0
|
||
|
|
||
|
if order == 'F' == get_best_order(&dst, ndim):
|
||
|
# see if both slices have Fortran order, transpose them to match our
|
||
|
# C-style indexing order
|
||
|
transpose_memslice(&src)
|
||
|
transpose_memslice(&dst)
|
||
|
|
||
|
refcount_copying(&dst, dtype_is_object, ndim, False)
|
||
|
copy_strided_to_strided(&src, &dst, ndim, itemsize)
|
||
|
refcount_copying(&dst, dtype_is_object, ndim, True)
|
||
|
|
||
|
free(tmpdata)
|
||
|
return 0
|
||
|
|
||
|
@cname('__pyx_memoryview_broadcast_leading')
|
||
|
cdef void broadcast_leading({{memviewslice_name}} *mslice,
|
||
|
int ndim,
|
||
|
int ndim_other) nogil:
|
||
|
cdef int i
|
||
|
cdef int offset = ndim_other - ndim
|
||
|
|
||
|
for i in range(ndim - 1, -1, -1):
|
||
|
mslice.shape[i + offset] = mslice.shape[i]
|
||
|
mslice.strides[i + offset] = mslice.strides[i]
|
||
|
mslice.suboffsets[i + offset] = mslice.suboffsets[i]
|
||
|
|
||
|
for i in range(offset):
|
||
|
mslice.shape[i] = 1
|
||
|
mslice.strides[i] = mslice.strides[0]
|
||
|
mslice.suboffsets[i] = -1
|
||
|
|
||
|
#
|
||
|
### Take care of refcounting the objects in slices. Do this separately from any copying,
|
||
|
### to minimize acquiring the GIL
|
||
|
#
|
||
|
|
||
|
@cname('__pyx_memoryview_refcount_copying')
|
||
|
cdef void refcount_copying({{memviewslice_name}} *dst, bint dtype_is_object,
|
||
|
int ndim, bint inc) nogil:
|
||
|
# incref or decref the objects in the destination slice if the dtype is
|
||
|
# object
|
||
|
if dtype_is_object:
|
||
|
refcount_objects_in_slice_with_gil(dst.data, dst.shape,
|
||
|
dst.strides, ndim, inc)
|
||
|
|
||
|
@cname('__pyx_memoryview_refcount_objects_in_slice_with_gil')
|
||
|
cdef void refcount_objects_in_slice_with_gil(char *data, Py_ssize_t *shape,
|
||
|
Py_ssize_t *strides, int ndim,
|
||
|
bint inc) with gil:
|
||
|
refcount_objects_in_slice(data, shape, strides, ndim, inc)
|
||
|
|
||
|
@cname('__pyx_memoryview_refcount_objects_in_slice')
|
||
|
cdef void refcount_objects_in_slice(char *data, Py_ssize_t *shape,
|
||
|
Py_ssize_t *strides, int ndim, bint inc):
|
||
|
cdef Py_ssize_t i
|
||
|
|
||
|
for i in range(shape[0]):
|
||
|
if ndim == 1:
|
||
|
if inc:
|
||
|
Py_INCREF((<PyObject **> data)[0])
|
||
|
else:
|
||
|
Py_DECREF((<PyObject **> data)[0])
|
||
|
else:
|
||
|
refcount_objects_in_slice(data, shape + 1, strides + 1,
|
||
|
ndim - 1, inc)
|
||
|
|
||
|
data += strides[0]
|
||
|
|
||
|
#
|
||
|
### Scalar to slice assignment
|
||
|
#
|
||
|
@cname('__pyx_memoryview_slice_assign_scalar')
|
||
|
cdef void slice_assign_scalar({{memviewslice_name}} *dst, int ndim,
|
||
|
size_t itemsize, void *item,
|
||
|
bint dtype_is_object) nogil:
|
||
|
refcount_copying(dst, dtype_is_object, ndim, False)
|
||
|
_slice_assign_scalar(dst.data, dst.shape, dst.strides, ndim,
|
||
|
itemsize, item)
|
||
|
refcount_copying(dst, dtype_is_object, ndim, True)
|
||
|
|
||
|
|
||
|
@cname('__pyx_memoryview__slice_assign_scalar')
|
||
|
cdef void _slice_assign_scalar(char *data, Py_ssize_t *shape,
|
||
|
Py_ssize_t *strides, int ndim,
|
||
|
size_t itemsize, void *item) nogil:
|
||
|
cdef Py_ssize_t i
|
||
|
cdef Py_ssize_t stride = strides[0]
|
||
|
cdef Py_ssize_t extent = shape[0]
|
||
|
|
||
|
if ndim == 1:
|
||
|
for i in range(extent):
|
||
|
memcpy(data, item, itemsize)
|
||
|
data += stride
|
||
|
else:
|
||
|
for i in range(extent):
|
||
|
_slice_assign_scalar(data, shape + 1, strides + 1,
|
||
|
ndim - 1, itemsize, item)
|
||
|
data += stride
|
||
|
|
||
|
|
||
|
############### BufferFormatFromTypeInfo ###############
|
||
|
cdef extern from *:
|
||
|
ctypedef struct __Pyx_StructField
|
||
|
|
||
|
cdef enum:
|
||
|
__PYX_BUF_FLAGS_PACKED_STRUCT
|
||
|
__PYX_BUF_FLAGS_INTEGER_COMPLEX
|
||
|
|
||
|
ctypedef struct __Pyx_TypeInfo:
|
||
|
char* name
|
||
|
__Pyx_StructField* fields
|
||
|
size_t size
|
||
|
size_t arraysize[8]
|
||
|
int ndim
|
||
|
char typegroup
|
||
|
char is_unsigned
|
||
|
int flags
|
||
|
|
||
|
ctypedef struct __Pyx_StructField:
|
||
|
__Pyx_TypeInfo* type
|
||
|
char* name
|
||
|
size_t offset
|
||
|
|
||
|
ctypedef struct __Pyx_BufFmt_StackElem:
|
||
|
__Pyx_StructField* field
|
||
|
size_t parent_offset
|
||
|
|
||
|
#ctypedef struct __Pyx_BufFmt_Context:
|
||
|
# __Pyx_StructField root
|
||
|
__Pyx_BufFmt_StackElem* head
|
||
|
|
||
|
struct __pyx_typeinfo_string:
|
||
|
char string[3]
|
||
|
|
||
|
__pyx_typeinfo_string __Pyx_TypeInfoToFormat(__Pyx_TypeInfo *)
|
||
|
|
||
|
|
||
|
@cname('__pyx_format_from_typeinfo')
|
||
|
cdef bytes format_from_typeinfo(__Pyx_TypeInfo *type):
|
||
|
cdef __Pyx_StructField *field
|
||
|
cdef __pyx_typeinfo_string fmt
|
||
|
cdef bytes part, result
|
||
|
|
||
|
if type.typegroup == 'S':
|
||
|
assert type.fields != NULL and type.fields.type != NULL
|
||
|
|
||
|
if type.flags & __PYX_BUF_FLAGS_PACKED_STRUCT:
|
||
|
alignment = b'^'
|
||
|
else:
|
||
|
alignment = b''
|
||
|
|
||
|
parts = [b"T{"]
|
||
|
field = type.fields
|
||
|
|
||
|
while field.type:
|
||
|
part = format_from_typeinfo(field.type)
|
||
|
parts.append(part + b':' + field.name + b':')
|
||
|
field += 1
|
||
|
|
||
|
result = alignment.join(parts) + b'}'
|
||
|
else:
|
||
|
fmt = __Pyx_TypeInfoToFormat(type)
|
||
|
if type.arraysize[0]:
|
||
|
extents = [unicode(type.arraysize[i]) for i in range(type.ndim)]
|
||
|
result = (u"(%s)" % u','.join(extents)).encode('ascii') + fmt.string
|
||
|
else:
|
||
|
result = fmt.string
|
||
|
|
||
|
return result
|