Source code for stream_framework.utils

from stream_framework.exceptions import DuplicateActivityException
import collections
from datetime import datetime, timedelta
import functools
import itertools
import logging
import six


logger = logging.getLogger(__name__)

MISSING = object()


[docs]class LRUCache: def __init__(self, capacity): self.capacity = capacity self.cache = collections.OrderedDict()
[docs] def get(self, key): try: value = self.cache.pop(key) self.cache[key] = value return value except KeyError: return MISSING
[docs] def set(self, key, value): try: self.cache.pop(key) except KeyError: if len(self.cache) >= self.capacity: self.cache.popitem(last=False) self.cache[key] = value
[docs]def chunks(iterable, n=10000): it = iter(iterable) while True: chunk = tuple(itertools.islice(it, n)) if not chunk: return yield chunk
epoch = datetime(1970, 1, 1)
[docs]def datetime_to_epoch(dt): ''' Convert datetime object to epoch with millisecond accuracy ''' delta = dt - epoch since_epoch = delta.total_seconds() return since_epoch
[docs]def epoch_to_datetime(time_): return epoch + timedelta(seconds=time_)
[docs]def make_list_unique(sequence, marker_function=None): ''' Makes items in a list unique Performance based on this blog post: http://www.peterbe.com/plog/uniqifiers-benchmark ''' seen = {} result = [] for item in sequence: # gets the marker marker = item if marker_function is not None: marker = marker_function(item) # if no longer unique make unique if marker in seen: continue seen[marker] = True result.append(item) return result
[docs]def warn_on_error(f, exceptions): import sys assert exceptions assert isinstance(exceptions, tuple) @functools.wraps(f) def wrapper(*args, **kwargs): try: return f(*args, **kwargs) except exceptions as e: logger.warn(six.text_type(e), exc_info=sys.exc_info(), extra={ 'data': { 'body': six.text_type(e), } }) return wrapper
[docs]def warn_on_duplicate(f): exceptions = (DuplicateActivityException,) return warn_on_error(f, exceptions)
[docs]class memoized(object): '''Decorator. Caches a function's return value each time it is called. If called later with the same arguments, the cached value is returned (not reevaluated). ''' def __init__(self, func): self.func = func self.cache = LRUCache(10000) def __call__(self, *args): if not isinstance(args, collections.Hashable): # uncacheable. a list, for instance. # better to not cache than blow up. return self.func(*args) if self.cache.get(args) is not MISSING: return self.cache.get(args) else: value = self.func(*args) self.cache.set(args, value) return value def __repr__(self): '''Return the function's docstring.''' return self.func.__doc__ def __get__(self, obj, objtype): '''Support instance methods.''' return functools.partial(self.__call__, obj)
[docs]def get_metrics_instance(): """ Returns an instance of the metric class as defined in stream_framework settings. """ from stream_framework import settings metric_cls = get_class_from_string(settings.STREAM_METRIC_CLASS) return metric_cls(**settings.STREAM_METRICS_OPTIONS)
[docs]def get_class_from_string(path, default=None): """ Return the class specified by the string. """ try: from importlib import import_module except ImportError: from django.utils.importlib import import_module i = path.rfind('.') module, attr = path[:i], path[i + 1:] mod = import_module(module) try: return getattr(mod, attr) except AttributeError: if default: return default else: raise ImportError( 'Cannot import name {} (from {})'.format(attr, mod))