Simple (but powerful) Caching Tools.
There are many libraries out there to deal with the same problem that this library tries to solve, but most of them fall short on at least one of the following points:
- Minimalism.
- Providing proper type hints to ease the user's life when using the library.
- Providing out-of-the-box support for asynchronous functions.
- Simplicity:
- In
power_cache
, capacity is measured just counting items, and not using their size as other libraries do. There are legitimate reasons to avoid the "sizes approach": it decreases performance, and it's highly error prone.
- In
- Extra flexibility:
power_cache
allows its decorators to not cache certain chosen values. This can be quite handy for some nullish/emptyish results. - Correctness:
- Some popular implementations incorrectly implement
__eq__
by just comparing object hashes. - Some popular implementations implement
__hash__
in a way that collisions will be more frequent than desirable.
- Some popular implementations incorrectly implement
- Performance: Even though
power_cache
is not outstanding in any sense when it comes to performance, at least it avoids some questionable decisions made in other libraries (like relying ondatetime
instead ofmonotonic
forttl
caches).
from power_cache import LRUCache
cache = LRUCache(capacity=3)
# We can also specify key & value types if we are using `mypy` or `pytypes`
cache = LRUCache[str, int](capacity=3)
cache['the answer to everything'] = 42
cache['the answer to everything'] # returns 42
cache['a'] = 1
cache['b'] = 2
cache['c'] = 3
# Raises KeyError, because the key was the least recently used, and the capacity
# is only 3, so the previous value was evicted.
cache['the answer to everything']
TTLCache
is very similar to LRUCache
, with the distinction that it marks
values as expired if they are too old.
from time import sleep
from power_cache import TTLCache
cache = TTLCache(capacity=3, ttl=120) # Values valid only for 2 minutes
# We can also specify key & value types if we are using `mypy` or `pytypes`
cache = TTLCache[str, int](capacity=3, ttl=120)
cache['the answer to everything'] = 42
cache['the answer to everything'] # returns 42
cache['a'] = 1
cache['b'] = 2
cache['c'] = 3
# Raises KeyError, because the key was the least recently used, and the capacity
# is only 3, so the previous value was evicted.
cache['the answer to everything']
assert len(cache) == 3
cache.evict_expired() # We can manually evict all expired values
assert len(cache) == 3 # Nothing was evicted because values are too recent
sleep(121)
# Now all values are marked as expired, but not evicted automatically, because
# that would spend too much CPU time.
assert len(cache) == 3
cache.evict_expired() # We can manually evict all expired values
assert len(cache) == 0
from power_cache import Memoize
# Runtime annotations are preserved.
# `capacity` must be always specified, while `cache_type` is "lru" by default.
@Memoize(capacity=3, cache_type="lru")
def my_function(): ...
@Memoize(capacity=3, cache_type="ttl", ttl=120)
def another_function(): ...
# We can instruct our memoizer to not save certain results, like `None`
@Memoize(capacity=3, results_to_discard=(None,))
def my_function(): ...
from power_cache import AsyncMemoize
# Runtime annotations are preserved.
# `capacity` must be always specified, while `cache_type` is "lru" by default.
@AsyncMemoize(capacity=3, cache_type="lru")
async def my_function(): ...
@AsyncMemoize(capacity=3, cache_type="ttl", ttl=120)
async def another_function(): ...
# We can instruct our memoizer to not save certain results, like `None`
@AsyncMemoize(capacity=3, results_to_discard=(None,))
def my_function(): ...