"Django-compatible disk and file backed cache." from functools import wraps from django.core.cache.backends.base import BaseCache try: from django.core.cache.backends.base import DEFAULT_TIMEOUT except ImportError: # For older versions of Django simply use 300 seconds. DEFAULT_TIMEOUT = 300 from .fanout import FanoutCache MARK = object() class DjangoCache(BaseCache): "Django-compatible disk and file backed cache." def __init__(self, directory, params): """Initialize DjangoCache instance. :param str directory: cache directory :param dict params: cache parameters """ super(DjangoCache, self).__init__(params) shards = params.get('SHARDS', 8) timeout = params.get('DATABASE_TIMEOUT', 0.010) options = params.get('OPTIONS', {}) self._directory = directory self._cache = FanoutCache(directory, shards, timeout, **options) @property def cache(self): "FanoutCache used by DjangoCache." return self._cache @property def directory(self): """Cache directory.""" return self._directory def deque(self, name): """Return Deque with given `name` in subdirectory. :param str name: subdirectory name for Deque :return: Deque with given name """ return self._cache.deque(name) def index(self, name): """Return Index with given `name` in subdirectory. :param str name: subdirectory name for Index :return: Index with given name """ return self._cache.index(name) def add(self, key, value, timeout=DEFAULT_TIMEOUT, version=None, read=False, tag=None, retry=True): """Set a value in the cache if the key does not already exist. If timeout is given, that timeout will be used for the key; otherwise the default cache timeout will be used. Return True if the value was stored, False otherwise. :param key: key for item :param value: value for item :param float timeout: seconds until the item expires (default 300 seconds) :param int version: key version number (default None, cache parameter) :param bool read: read value as bytes from file (default False) :param str tag: text to associate with key (default None) :param bool retry: retry if database timeout occurs (default True) :return: True if item was added """ # pylint: disable=arguments-differ key = self.make_key(key, version=version) timeout = self.get_backend_timeout(timeout=timeout) return self._cache.add(key, value, timeout, read, tag, retry) def get(self, key, default=None, version=None, read=False, expire_time=False, tag=False, retry=False): """Fetch a given key from the cache. If the key does not exist, return default, which itself defaults to None. :param key: key for item :param default: return value if key is missing (default None) :param int version: key version number (default None, cache parameter) :param bool read: if True, return file handle to value (default False) :param float expire_time: if True, return expire_time in tuple (default False) :param tag: if True, return tag in tuple (default False) :param bool retry: retry if database timeout occurs (default False) :return: value for item if key is found else default """ # pylint: disable=arguments-differ key = self.make_key(key, version=version) return self._cache.get(key, default, read, expire_time, tag, retry) def read(self, key, version=None): """Return file handle corresponding to `key` from Cache. :param key: Python key to retrieve :param int version: key version number (default None, cache parameter) :return: file open for reading in binary mode :raises KeyError: if key is not found """ key = self.make_key(key, version=version) return self._cache.read(key) def set(self, key, value, timeout=DEFAULT_TIMEOUT, version=None, read=False, tag=None, retry=True): """Set a value in the cache. If timeout is given, that timeout will be used for the key; otherwise the default cache timeout will be used. :param key: key for item :param value: value for item :param float timeout: seconds until the item expires (default 300 seconds) :param int version: key version number (default None, cache parameter) :param bool read: read value as bytes from file (default False) :param str tag: text to associate with key (default None) :param bool retry: retry if database timeout occurs (default True) :return: True if item was set """ # pylint: disable=arguments-differ key = self.make_key(key, version=version) timeout = self.get_backend_timeout(timeout=timeout) return self._cache.set(key, value, timeout, read, tag, retry) def touch(self, key, timeout=DEFAULT_TIMEOUT, version=None, retry=True): """Touch a key in the cache. If timeout is given, that timeout will be used for the key; otherwise the default cache timeout will be used. :param key: key for item :param float timeout: seconds until the item expires (default 300 seconds) :param int version: key version number (default None, cache parameter) :param bool retry: retry if database timeout occurs (default True) :return: True if key was touched """ # pylint: disable=arguments-differ key = self.make_key(key, version=version) timeout = self.get_backend_timeout(timeout=timeout) return self._cache.touch(key, timeout, retry) def pop(self, key, default=None, version=None, expire_time=False, tag=False, retry=True): """Remove corresponding item for `key` from cache and return value. If `key` is missing, return `default`. Operation is atomic. Concurrent operations will be serialized. :param key: key for item :param default: return value if key is missing (default None) :param int version: key version number (default None, cache parameter) :param float expire_time: if True, return expire_time in tuple (default False) :param tag: if True, return tag in tuple (default False) :param bool retry: retry if database timeout occurs (default True) :return: value for item if key is found else default """ key = self.make_key(key, version=version) return self._cache.pop(key, default, expire_time, tag, retry) def delete(self, key, version=None, retry=True): """Delete a key from the cache, failing silently. :param key: key for item :param int version: key version number (default None, cache parameter) :param bool retry: retry if database timeout occurs (default True) :return: True if item was deleted """ # pylint: disable=arguments-differ key = self.make_key(key, version=version) self._cache.delete(key, retry) def incr(self, key, delta=1, version=None, default=None, retry=True): """Increment value by delta for item with key. If key is missing and default is None then raise KeyError. Else if key is missing and default is not None then use default for value. Operation is atomic. All concurrent increment operations will be counted individually. Assumes value may be stored in a SQLite column. Most builds that target machines with 64-bit pointer widths will support 64-bit signed integers. :param key: key for item :param int delta: amount to increment (default 1) :param int version: key version number (default None, cache parameter) :param int default: value if key is missing (default None) :param bool retry: retry if database timeout occurs (default True) :return: new value for item on success else None :raises ValueError: if key is not found and default is None """ # pylint: disable=arguments-differ key = self.make_key(key, version=version) try: return self._cache.incr(key, delta, default, retry) except KeyError: raise ValueError("Key '%s' not found" % key) def decr(self, key, delta=1, version=None, default=None, retry=True): """Decrement value by delta for item with key. If key is missing and default is None then raise KeyError. Else if key is missing and default is not None then use default for value. Operation is atomic. All concurrent decrement operations will be counted individually. Unlike Memcached, negative values are supported. Value may be decremented below zero. Assumes value may be stored in a SQLite column. Most builds that target machines with 64-bit pointer widths will support 64-bit signed integers. :param key: key for item :param int delta: amount to decrement (default 1) :param int version: key version number (default None, cache parameter) :param int default: value if key is missing (default None) :param bool retry: retry if database timeout occurs (default True) :return: new value for item on success else None :raises ValueError: if key is not found and default is None """ # pylint: disable=arguments-differ return self.incr(key, -delta, version, default, retry) def has_key(self, key, version=None): """Returns True if the key is in the cache and has not expired. :param key: key for item :param int version: key version number (default None, cache parameter) :return: True if key is found """ key = self.make_key(key, version=version) return key in self._cache def expire(self): """Remove expired items from cache. :return: count of items removed """ return self._cache.expire() def stats(self, enable=True, reset=False): """Return cache statistics hits and misses. :param bool enable: enable collecting statistics (default True) :param bool reset: reset hits and misses to 0 (default False) :return: (hits, misses) """ return self._cache.stats(enable=enable, reset=reset) def create_tag_index(self): """Create tag index on cache database. It is better to initialize cache with `tag_index=True` than use this. :raises Timeout: if database timeout occurs """ self._cache.create_tag_index() def drop_tag_index(self): """Drop tag index on cache database. :raises Timeout: if database timeout occurs """ self._cache.drop_tag_index() def evict(self, tag): """Remove items with matching `tag` from cache. :param str tag: tag identifying items :return: count of items removed """ return self._cache.evict(tag) def cull(self): """Cull items from cache until volume is less than size limit. :return: count of items removed """ return self._cache.cull() def clear(self): "Remove *all* values from the cache at once." return self._cache.clear() def close(self, **kwargs): "Close the cache connection." # pylint: disable=unused-argument self._cache.close() def get_backend_timeout(self, timeout=DEFAULT_TIMEOUT): """Return seconds to expiration. :param float timeout: seconds until the item expires (default 300 seconds) """ if timeout == DEFAULT_TIMEOUT: timeout = self.default_timeout elif timeout == 0: # ticket 21147 - avoid time.time() related precision issues timeout = -1 return None if timeout is None else timeout def memoize(self, name=None, timeout=DEFAULT_TIMEOUT, version=None, typed=False, tag=None): """Memoizing cache decorator. Decorator to wrap callable with memoizing function using cache. Repeated calls with the same arguments will lookup result in cache and avoid function evaluation. If name is set to None (default), the callable name will be determined automatically. If typed is set to True, function arguments of different types will be cached separately. For example, f(3) and f(3.0) will be treated as distinct calls with distinct results. The original underlying function is accessible through the __wrapped__ attribute. This is useful for introspection, for bypassing the cache, or for rewrapping the function with a different cache. Remember to call memoize when decorating a callable. If you forget, then a TypeError will occur. :param str name: name given for callable (default None, automatic) :param float timeout: seconds until the item expires (default 300 seconds) :param int version: key version number (default None, cache parameter) :param bool typed: cache different types separately (default False) :param str tag: text to associate with arguments (default None) :return: callable decorator """ # Caution: Nearly identical code exists in memo.memoize if callable(name): raise TypeError('name cannot be callable') def decorator(function): "Decorator created by memoize call for callable." if name is None: try: reference = function.__qualname__ except AttributeError: reference = function.__name__ reference = function.__module__ + reference else: reference = name reference = (reference,) @wraps(function) def wrapper(*args, **kwargs): "Wrapper for callable to cache arguments and return values." key = reference + args if kwargs: key += (MARK,) sorted_items = sorted(kwargs.items()) for item in sorted_items: key += item if typed: key += tuple(type(arg) for arg in args) if kwargs: key += tuple(type(value) for _, value in sorted_items) result = self.get(key, MARK, version, retry=True) if result is MARK: result = function(*args, **kwargs) self.set(key, result, timeout, version, tag=tag, retry=True) return result return wrapper return decorator