diff options
author | Kenneth Heafield <github@kheafield.com> | 2012-08-03 07:46:54 -0400 |
---|---|---|
committer | Kenneth Heafield <github@kheafield.com> | 2012-08-03 07:46:54 -0400 |
commit | 122f46c31102b683eaab3ad81a3a98accbc694bb (patch) | |
tree | 8d499d789b159ebed25bb23b6983813d064a6296 /python/src/vectors.pxi | |
parent | ac664bdb0e481539cf77098a7dd0e1ec8d937ba0 (diff) | |
parent | 193d137056c3c4f73d66f8db84691d63307de894 (diff) |
Merge branch 'master' of github.com:redpony/cdec
Diffstat (limited to 'python/src/vectors.pxi')
-rw-r--r-- | python/src/vectors.pxi | 145 |
1 files changed, 145 insertions, 0 deletions
diff --git a/python/src/vectors.pxi b/python/src/vectors.pxi new file mode 100644 index 00000000..989a6a7c --- /dev/null +++ b/python/src/vectors.pxi @@ -0,0 +1,145 @@ +from cython.operator cimport preincrement as pinc + +cdef class DenseVector: + cdef vector[weight_t]* vector + cdef bint owned # if True, do not manage memory + + def __init__(self): + self.vector = new vector[weight_t]() + self.owned = False + + def __dealloc__(self): + if not self.owned: + del self.vector + + def __len__(self): + return self.vector.size() + + def __getitem__(self, char* fname): + cdef int fid = FDConvert(fname) + if 0 <= fid < self.vector.size(): + return self.vector[0][fid] + raise KeyError(fname) + + def __setitem__(self, char* fname, float value): + cdef int fid = FDConvert(<char *>fname) + if fid < 0: raise KeyError(fname) + if self.vector.size() <= fid: + self.vector.resize(fid + 1) + self.vector[0][fid] = value + + def __iter__(self): + cdef unsigned fid + for fid in range(1, self.vector.size()): + yield FDConvert(fid).c_str(), self.vector[0][fid] + + def dot(self, SparseVector other): + return other.dot(self) + + def tosparse(self): + cdef SparseVector sparse = SparseVector.__new__(SparseVector) + sparse.vector = new FastSparseVector[weight_t]() + InitSparseVector(self.vector[0], sparse.vector) + return sparse + +cdef class SparseVector: + cdef FastSparseVector[weight_t]* vector + + def __init__(self): + self.vector = new FastSparseVector[weight_t]() + + def __dealloc__(self): + del self.vector + + def copy(self): + return self * 1 + + def __getitem__(self, char* fname): + cdef int fid = FDConvert(fname) + if fid < 0: raise KeyError(fname) + return self.vector.value(fid) + + def __setitem__(self, char* fname, float value): + cdef int fid = FDConvert(<char *>fname) + if fid < 0: raise KeyError(fname) + self.vector.set_value(fid, value) + + def __iter__(self): + cdef FastSparseVector[weight_t].const_iterator* it = new FastSparseVector[weight_t].const_iterator(self.vector[0], False) + cdef unsigned i + try: + for i in range(self.vector.size()): + yield (FDConvert(it[0].ptr().first).c_str(), it[0].ptr().second) + pinc(it[0]) # ++it + finally: + del it + + def dot(self, other): + if isinstance(other, DenseVector): + return self.vector.dot((<DenseVector> other).vector[0]) + elif isinstance(other, SparseVector): + return self.vector.dot((<SparseVector> other).vector[0]) + raise TypeError('cannot take the dot product of %s and SparseVector' % type(other)) + + def __richcmp__(SparseVector x, SparseVector y, int op): + if op == 2: # == + return x.vector[0] == y.vector[0] + elif op == 3: # != + return not (x == y) + raise NotImplemented('comparison not implemented for SparseVector') + + def __len__(self): + return self.vector.size() + + def __contains__(self, char* fname): + return self.vector.nonzero(FDConvert(fname)) + + def __neg__(self): + cdef SparseVector result = SparseVector.__new__(SparseVector) + result.vector = new FastSparseVector[weight_t](self.vector[0]) + result.vector[0] *= -1.0 + return result + + def __iadd__(SparseVector self, SparseVector other): + self.vector[0] += other.vector[0] + return self + + def __isub__(SparseVector self, SparseVector other): + self.vector[0] -= other.vector[0] + return self + + def __imul__(SparseVector self, float scalar): + self.vector[0] *= scalar + return self + + def __idiv__(SparseVector self, float scalar): + self.vector[0] /= scalar + return self + + def __add__(SparseVector x, SparseVector y): + cdef SparseVector result = SparseVector.__new__(SparseVector) + result.vector = new FastSparseVector[weight_t](x.vector[0] + y.vector[0]) + return result + + def __sub__(SparseVector x, SparseVector y): + cdef SparseVector result = SparseVector.__new__(SparseVector) + result.vector = new FastSparseVector[weight_t](x.vector[0] - y.vector[0]) + return result + + def __mul__(x, y): + cdef SparseVector vector + cdef float scalar + if isinstance(x, SparseVector): vector, scalar = x, y + else: vector, scalar = y, x + cdef SparseVector result = SparseVector.__new__(SparseVector) + result.vector = new FastSparseVector[weight_t](vector.vector[0] * scalar) + return result + + def __div__(x, y): + cdef SparseVector vector + cdef float scalar + if isinstance(x, SparseVector): vector, scalar = x, y + else: vector, scalar = y, x + cdef SparseVector result = SparseVector.__new__(SparseVector) + result.vector = new FastSparseVector[weight_t](vector.vector[0] / scalar) + return result |