# cython: infer_types=True # coding: utf8 from __future__ import unicode_literals, absolute_import cimport cython from libc.string cimport memcpy from libcpp.set cimport set from libc.stdint cimport uint32_t from murmurhash.mrmr cimport hash64, hash32 import ujson from .symbols import IDS as SYMBOLS_BY_STR from .symbols import NAMES as SYMBOLS_BY_INT from .typedefs cimport hash_t from .compat import json_dumps from .errors import Errors from . import util cpdef hash_t hash_string(unicode string) except 0: chars = string.encode('utf8') return hash_utf8(chars, len(chars)) cdef hash_t hash_utf8(char* utf8_string, int length) nogil: return hash64(utf8_string, length, 1) cdef uint32_t hash32_utf8(char* utf8_string, int length) nogil: return hash32(utf8_string, length, 1) cdef unicode decode_Utf8Str(const Utf8Str* string): cdef int i, length if string.s[0] < sizeof(string.s) and string.s[0] != 0: return string.s[1:string.s[0]+1].decode('utf8') elif string.p[0] < 255: return string.p[1:string.p[0]+1].decode('utf8') else: i = 0 length = 0 while string.p[i] == 255: i += 1 length += 255 length += string.p[i] i += 1 return string.p[i:length + i].decode('utf8') cdef Utf8Str* _allocate(Pool mem, const unsigned char* chars, uint32_t length) except *: cdef int n_length_bytes cdef int i cdef Utf8Str* string = mem.alloc(1, sizeof(Utf8Str)) cdef uint32_t ulength = length if length < sizeof(string.s): string.s[0] = length memcpy(&string.s[1], chars, length) return string elif length < 255: string.p = mem.alloc(length + 1, sizeof(unsigned char)) string.p[0] = length memcpy(&string.p[1], chars, length) return string else: i = 0 n_length_bytes = (length // 255) + 1 string.p = mem.alloc(length + n_length_bytes, sizeof(unsigned char)) for i in range(n_length_bytes-1): string.p[i] = 255 string.p[n_length_bytes-1] = length % 255 memcpy(&string.p[n_length_bytes], chars, length) return string cdef class StringStore: """Look up strings by 64-bit hashes.""" def __init__(self, strings=None, freeze=False): """Create the StringStore. strings (iterable): A sequence of unicode strings to add to the store. RETURNS (StringStore): The newly constructed object. """ self.mem = Pool() self._map = PreshMap() if strings is not None: for string in strings: self.add(string) def __getitem__(self, object string_or_id): """Retrieve a string from a given hash, or vice versa. string_or_id (bytes, unicode or uint64): The value to encode. Returns (unicode or uint64): The value to be retrieved. """ if isinstance(string_or_id, basestring) and len(string_or_id) == 0: return 0 elif string_or_id == 0: return u'' elif string_or_id in SYMBOLS_BY_STR: return SYMBOLS_BY_STR[string_or_id] cdef hash_t key if isinstance(string_or_id, unicode): key = hash_string(string_or_id) return key elif isinstance(string_or_id, bytes): key = hash_utf8(string_or_id, len(string_or_id)) return key elif string_or_id < len(SYMBOLS_BY_INT): return SYMBOLS_BY_INT[string_or_id] else: key = string_or_id self.hits.insert(key) utf8str = self._map.get(key) if utf8str is NULL: raise KeyError(Errors.E018.format(hash_value=string_or_id)) else: return decode_Utf8Str(utf8str) def add(self, string): """Add a string to the StringStore. string (unicode): The string to add. RETURNS (uint64): The string's hash value. """ if isinstance(string, unicode): if string in SYMBOLS_BY_STR: return SYMBOLS_BY_STR[string] key = hash_string(string) self.intern_unicode(string) elif isinstance(string, bytes): if string in SYMBOLS_BY_STR: return SYMBOLS_BY_STR[string] key = hash_utf8(string, len(string)) self._intern_utf8(string, len(string)) else: raise TypeError(Errors.E017.format(value_type=type(string))) return key def __len__(self): """The number of strings in the store. RETURNS (int): The number of strings in the store. """ return self.keys.size() def __contains__(self, string not None): """Check whether a string is in the store. string (unicode): The string to check. RETURNS (bool): Whether the store contains the string. """ cdef hash_t key if isinstance(string, int) or isinstance(string, long): if string == 0: return True key = string elif len(string) == 0: return True elif string in SYMBOLS_BY_STR: return True elif isinstance(string, unicode): key = hash_string(string) else: string = string.encode('utf8') key = hash_utf8(string, len(string)) if key < len(SYMBOLS_BY_INT): return True else: self.hits.insert(key) return self._map.get(key) is not NULL def __iter__(self): """Iterate over the strings in the store, in order. YIELDS (unicode): A string in the store. """ cdef int i cdef hash_t key for i in range(self.keys.size()): key = self.keys[i] self.hits.insert(key) utf8str = self._map.get(key) yield decode_Utf8Str(utf8str) # TODO: Iterate OOV here? def __reduce__(self): strings = list(self) return (StringStore, (strings,), None, None, None) def to_disk(self, path): """Save the current state to a directory. path (unicode or Path): A path to a directory, which will be created if it doesn't exist. Paths may be either strings or Path-like objects. """ path = util.ensure_path(path) strings = list(self) with path.open('w') as file_: file_.write(json_dumps(strings)) def from_disk(self, path): """Loads state from a directory. Modifies the object in place and returns it. path (unicode or Path): A path to a directory. Paths may be either strings or `Path`-like objects. RETURNS (StringStore): The modified `StringStore` object. """ path = util.ensure_path(path) with path.open('r') as file_: strings = ujson.load(file_) prev = list(self) self._reset_and_load(strings) for word in prev: self.add(word) return self def to_bytes(self, **exclude): """Serialize the current state to a binary string. **exclude: Named attributes to prevent from being serialized. RETURNS (bytes): The serialized form of the `StringStore` object. """ return json_dumps(list(self)) def from_bytes(self, bytes_data, **exclude): """Load state from a binary string. bytes_data (bytes): The data to load from. **exclude: Named attributes to prevent from being loaded. RETURNS (StringStore): The `StringStore` object. """ strings = ujson.loads(bytes_data) prev = list(self) self._reset_and_load(strings) for word in prev: self.add(word) return self def _reset_and_load(self, strings): self.mem = Pool() self._map = PreshMap() self.keys.clear() self.hits.clear() for string in strings: self.add(string) def _cleanup_stale_strings(self, excepted): """ excepted (list): Strings that should not be removed. RETURNS (keys, strings): Dropped strings and keys that can be dropped from other places """ if self.hits.size() == 0: # If we don't have any hits, just skip cleanup return cdef vector[hash_t] tmp dropped_strings = [] dropped_keys = [] for i in range(self.keys.size()): key = self.keys[i] # Here we cannot use __getitem__ because it also set hit. utf8str = self._map.get(key) value = decode_Utf8Str(utf8str) if self.hits.count(key) != 0 or value in excepted: tmp.push_back(key) else: dropped_keys.append(key) dropped_strings.append(value) self.keys.swap(tmp) strings = list(self) self._reset_and_load(strings) # Here we have strings but hits to it should be reseted self.hits.clear() return dropped_keys, dropped_strings cdef const Utf8Str* intern_unicode(self, unicode py_string): # 0 means missing, but we don't bother offsetting the index. cdef bytes byte_string = py_string.encode('utf8') return self._intern_utf8(byte_string, len(byte_string)) @cython.final cdef const Utf8Str* _intern_utf8(self, char* utf8_string, int length): # TODO: This function's API/behaviour is an unholy mess... # 0 means missing, but we don't bother offsetting the index. cdef hash_t key = hash_utf8(utf8_string, length) cdef Utf8Str* value = self._map.get(key) if value is not NULL: return value value = _allocate(self.mem, utf8_string, length) self._map.set(key, value) self.hits.insert(key) self.keys.push_back(key) return value