Metadata-Version: 2.4
Name: atomic-lru
Version: 0.0.5
Summary: Thread-safe, dependency-free in-memory LRU storage library for Python 3.10+ (with optional TTL)
Author-email: Fabien MARTY <fabien.marty@gmail.com>
Maintainer-email: Fabien MARTY <fab@fabien-marty.dev>
License-Expression: MIT
Project-URL: Homepage, https://fabien-marty.github.io/atomic-lru
Project-URL: Repository, https://github.com/fabien-marty/atomic-lru.git
Project-URL: Documentation, https://fabien-marty.github.io/atomic-lru
Requires-Python: >=3.10
Description-Content-Type: text/markdown
License-File: LICENSE
Dynamic: license-file

<!-- *** GENERATED FILE - DO NOT EDIT *** -->
# Atomic LRU

## What is this?

This is a **thread-safe** and **dependency-free** **in-memory** **LRU storage** Python 3.10+ library with optional Time To Live (TTL).

You can define:

- **limits** (`max-items` or `max-size-in-bytes`)
- **TTL expiration** (globally or per item)

to prevent the storage from growing too big.

You will get an automatic **LRU eviction** of the least recently used items when the limits are reached.

## Features

- Thread-Safe
- (optional) TTL expiration *(globally or per item)*
- (optional) Total size limit *(in bytes)* [^1]
- (optional) Max items limit
- Automatic LRU eviction *(when the limits are reached)*
- Full-typing support
- High level `Cache` API **with** automatic serialization/deserialization [^2]
- Low level `Storage` API **without** serialization/deserialization *(store only references to given objects)*

## Quickstart

### Installation

`pip install atomic-lru` (or equivalent for your package manager)

### High level API *(with automatic serialization/deserialization)*

The main use-case is to use it as a **cache** for your data. You store any kind of data type which will be **automatically serialized** to bytes. [^2]

```python
from atomic_lru import CACHE_MISS, Cache

# Create a Cache object instance (with a size limit of 1MB)
# (this object is thread-safe, so you can use it from multiple threads)
cache = Cache(size_limit_in_bytes=1_000_000, default_ttl=3600)

# Let's store something (a dictionnary here) in the cache with a custom TTL
cache.set(key="user:123", value={"name": "Alice", "age": 30}, ttl=60)

# ...

# Let's retrieve it
user = cache.get(key="user:123")

if user is not CACHE_MISS:
    # cache hit
    print(user["name"])
```

### Low level API *(without serialization/deserialization)*

But you can use it at a lower level to store any kind of data type **without serialization**. In that case, you will loose the `max-size-in-bytes` feature but you still get the `max-items` feature.

```python
from atomic_lru import CACHE_MISS, Storage


class ExpensiveObject:
    """An expensive object that is not serializable."""

    pass


# Create a Storage object instance to store ExpensiveObject instances
# (this object is thread-safe, so you can use it from multiple threads)
storage = Storage[ExpensiveObject](max_items=100, default_ttl=3600)

# Create and store an ExpensiveObject instance
value = ExpensiveObject()
storage.set("key1", value, ttl=60)

# ...

# Let's retrieve it
obj = storage.get("key1")

if obj is not CACHE_MISS:
    # cache hit
    assert isinstance(obj, ExpensiveObject)
    assert id(obj) == id(value)  # this is the same object instance
```

## Full API reference

Refer to the [API reference](reference/api.md) for the full API.

## DEV

This library is managed with `uv` and a `Makefile`. Execute:

- `uv sync` to create the virtual environment
- `make lint` to lint the code (style, checks, types, architecture) and fix obvious things
- `make test` to execute unit tests
- `make doc` to generate the documentation

See https://docs.astral.sh/uv/getting-started/installation/ to install `uv` if you need to.

[^1]: This feature is only available when using the high level `Cache` API.
[^2]: By default, `pickle` is used for serialization/deserialization but you can provide your own serializer/deserializer if you want to use a different format.
