Metadata-Version: 2.4
Name: cachestack
Version: 0.1.0
Summary: A modern, async-first, multi-backend Python caching library.
Project-URL: Homepage, https://github.com/sarthakbhatkar/cachestack
Project-URL: Repository, https://github.com/sarthakbhatkar/cachestack
Project-URL: Issues, https://github.com/sarthakbhatkar/cachestack/issues
Project-URL: Changelog, https://github.com/sarthakbhatkar/cachestack/blob/main/CHANGELOG.md
Author: Sarthak Bhatkar
License: MIT
License-File: LICENSE
Keywords: async,asyncio,cache,caching,memcached,postgres,redis
Classifier: Development Status :: 4 - Beta
Classifier: Framework :: AsyncIO
Classifier: Intended Audience :: Developers
Classifier: License :: OSI Approved :: MIT License
Classifier: Programming Language :: Python :: 3
Classifier: Programming Language :: Python :: 3.11
Classifier: Programming Language :: Python :: 3.12
Classifier: Topic :: Database
Classifier: Topic :: Software Development :: Libraries :: Python Modules
Requires-Python: >=3.11
Provides-Extra: all
Requires-Dist: aiomcache>=0.8; extra == 'all'
Requires-Dist: asyncpg>=0.29; extra == 'all'
Requires-Dist: cachetools>=5.0; extra == 'all'
Requires-Dist: msgpack>=1.0; extra == 'all'
Requires-Dist: redis>=5.0; extra == 'all'
Provides-Extra: dev
Requires-Dist: cachetools>=5.0; extra == 'dev'
Requires-Dist: pytest-asyncio>=0.23; extra == 'dev'
Requires-Dist: pytest>=8.0; extra == 'dev'
Provides-Extra: memcached
Requires-Dist: aiomcache>=0.8; extra == 'memcached'
Provides-Extra: memory
Requires-Dist: cachetools>=5.0; extra == 'memory'
Provides-Extra: msgpack
Requires-Dist: msgpack>=1.0; extra == 'msgpack'
Provides-Extra: postgres
Requires-Dist: asyncpg>=0.29; extra == 'postgres'
Provides-Extra: redis
Requires-Dist: redis>=5.0; extra == 'redis'
Description-Content-Type: text/markdown

# ⚡ CacheStack

> Modern · Async-First · Multi-Backend Python Caching

![MIT License](https://img.shields.io/badge/License-MIT-green.svg)
![Python 3.11+](https://img.shields.io/badge/python-3.11+-blue.svg)
![BlackDuck Clean](https://img.shields.io/badge/BlackDuck-Clean-brightgreen.svg)
![Zero Required Dependencies](https://img.shields.io/badge/dependencies-zero%20required-purple.svg)

CacheStack is a modern Python caching library with a unified API across **Memory, File, Redis, PostgreSQL, and Memcached** backends. Built async-first, designed for production, and clean for enterprise security scans (BlackDuck, FOSSA, Snyk).

---

## ✨ Features

- **Unified API** — same interface across all backends
- **Async-native** — built for `asyncio` from the ground up
- **Tiered caching** — L1 memory → L2 file/Redis → L3 Postgres with automatic backfill
- **Stampede protection** — per-key async locks prevent thundering herd
- **Flexible decorators** — `@cached` and `@invalidate` for sync and async functions
- **Pluggable serializers** — Pickle (default), JSON, Msgpack — supported across all backends
- **File cache** — zero-dependency persistent cache, BlackDuck safe, atomic writes, Windows-safe
- **BlackDuck-friendly** — MIT license, clean SPDX metadata, no ambiguous dependencies
- **Observability** — per-backend and per-layer hit/miss/error stats

---

## 📦 Installation

### Core (no dependencies)
```bash
pip install cachestack
```

### With backends
```bash
pip install cachestack[memory]      # In-memory (LRU/LFU/TTL via cachetools)
pip install cachestack[redis]       # Redis backend
pip install cachestack[postgres]    # PostgreSQL via asyncpg
pip install cachestack[memcached]   # Memcached backend
pip install cachestack[msgpack]     # Msgpack serializer
pip install cachestack[all]         # Everything
```

> **Note:** `FileBackend` requires no extra install — it uses Python stdlib only.

---

## 🚀 Quick Start

### Memory Cache
```python
from cachestack import MemoryBackend, MemoryConfig

cache = MemoryBackend(MemoryConfig(policy="lru", maxsize=1024))

await cache.set("user:1", {"name": "Sarthak"}, ttl=300)
user = await cache.get("user:1")
```

### File Cache
```python
from cachestack import FileBackend, FileConfig

# Zero dependencies — pure Python stdlib, BlackDuck safe
# Supports any Python object via Pickle (default serializer)
cache = FileBackend(FileConfig(
    directory="./cache",
    ttl=3600,
    namespace="myapp",
))

# Cache any Python object — datetime, sets, custom classes all work
import datetime
await cache.set("ts", datetime.datetime.now())
await cache.set("report", report_data, ttl=86400)

# Proactively clean up expired files to reclaim disk space
deleted = await cache.purge_expired()

# Or enable automatic background purging every hour during writes
cache = FileBackend(FileConfig(
    directory="./cache",
    auto_purge_interval=3600,  # purge expired files every 3600s passively
))
```

### Redis Cache
```python
from cachestack import RedisBackend, RedisConfig

cache = RedisBackend(RedisConfig(
    dsn="redis://localhost:6379/0",
    ttl=600,
    namespace="myapp",
))

await cache.set("session:abc", {"user_id": 1})
session = await cache.get("session:abc")
```

### PostgreSQL Cache
```python
from cachestack import PostgresBackend, PostgresConfig

cache = PostgresBackend(PostgresConfig(
    dsn="postgresql://user:password@localhost/mydb",
    ttl=3600,
    table="cache",
))

# Table is auto-created on first use
await cache.set("report:q3", report_data, ttl=86400)
report = await cache.get("report:q3")

# Proactively purge expired entries
deleted = await cache.purge_expired()
```

### Memcached Cache
```python
from cachestack import MemcachedBackend, MemcachedConfig

cache = MemcachedBackend(MemcachedConfig(host="localhost", port=11211))

await cache.set("key", "value", ttl=60)
value = await cache.get("key")
```

---

## 🗂️ Tiered Caching

CacheStack's most powerful feature. Stack backends from fastest to slowest — reads check L1 first and automatically backfill faster layers on a miss.

```python
from cachestack import (
    TieredCache, WriteStrategy,
    MemoryBackend, MemoryConfig,
    FileBackend, FileConfig,
    RedisBackend, RedisConfig,
    PostgresBackend, PostgresConfig,
)

# 2-layer: Memory + File (zero infrastructure required)
cache = TieredCache([
    MemoryBackend(MemoryConfig(maxsize=512)),                  # L1: RAM
    FileBackend(FileConfig(directory="./cache", ttl=3600)),   # L2: Disk
])

# 3-layer: Memory + Redis + Postgres
cache = TieredCache(
    backends=[
        MemoryBackend(MemoryConfig(maxsize=512)),            # L1: RAM (fastest)
        RedisBackend(RedisConfig(dsn="redis://localhost")),  # L2: Redis
        PostgresBackend(PostgresConfig(dsn="postgresql://localhost/db")),  # L3: Postgres
    ],
    write_strategy=WriteStrategy.WRITE_THROUGH,  # or WRITE_BACK
)

await cache.set("key", "value", ttl=300)
value = await cache.get("key")  # Checks L1 → L2 → L3, backfills on miss
```

### Write Strategies

| Strategy | Behaviour | Best For |
|---|---|---|
| `WRITE_THROUGH` | Writes to all layers immediately | Read-heavy, consistency matters |
| `WRITE_BACK` | Writes to L1 only | Write-heavy, eventual consistency |

---

## 🎨 Decorators

### `@cached` — Cache function return values

Works on both **sync and async** functions.

```python
from cachestack import cached, MemoryBackend

cache = MemoryBackend()

# Basic usage
@cached(cache=cache, ttl=60)
async def get_user(user_id: int):
    return await db.fetch(user_id)

# Custom key builder + condition
@cached(
    cache=cache,
    ttl=300,
    key_builder=lambda fn, args, kw: f"user:{args[0]}",
    condition=lambda v: v is not None,  # only cache non-None results
)
async def get_profile(user_id: int):
    ...

# Works on sync functions too
@cached(cache=cache, ttl=60)
def compute_expensive(x: int):
    return x ** 3
```

### `@invalidate` — Bust the cache on writes

```python
from cachestack import invalidate

key_fn = lambda fn, args, kw: f"user:{args[0]}"

@cached(cache=cache, ttl=300, key_builder=key_fn)
async def get_user(user_id: int):
    ...

@invalidate(cache=cache, key_builder=key_fn)
async def update_user(user_id: int, data: dict):
    await db.update(user_id, data)  # Cache auto-invalidated after this
```

---

## 🔧 Serializers

All backends support pluggable serializers. `FileBackend` defaults to `PickleSerializer` so any Python object can be cached safely. Network backends (Redis, Memcached) also default to Pickle.

| Serializer | Best For | Type Support | Speed |
|---|---|---|---|
| `PickleSerializer` *(default)* | Any Python object | All picklable types (datetime, set, custom classes) | Fast |
| `JsonSerializer` | Human-readable files / APIs | str, int, list, dict, bool | Medium |
| `MsgpackSerializer` | High-throughput systems | Most primitive types | Fastest |

```python
from cachestack import FileBackend, FileConfig, RedisBackend, RedisConfig
from cachestack import JsonSerializer, MsgpackSerializer

# File cache with JSON — human-readable files on disk
cache = FileBackend(FileConfig(directory="./cache", serializer=JsonSerializer()))

# Redis with Msgpack — compact binary, fast for high throughput
cache = RedisBackend(RedisConfig(serializer=MsgpackSerializer()))  # needs cachestack[msgpack]
```

---

## 📊 Observability

Every backend exposes a `.stats()` method.

```python
stats = await cache.stats()
# Memory backend:
# {
#   "backend": "memory",
#   "policy": "lru",
#   "hits": 142,
#   "misses": 8,
#   "errors": 0,
#   "size": 58,
#   "maxsize": 1024
# }

# File backend:
# {
#   "backend": "file",
#   "directory": "/abs/path/to/cache",
#   "namespace": "myapp",
#   "serializer": "PickleSerializer",
#   "hits": 98,
#   "misses": 12,
#   "errors": 0,
#   "files_on_disk": 43,
#   "writes": 110
# }

# TieredCache returns per-layer breakdown:
stats = await tiered.stats()
# {
#   "backend": "tiered",
#   "layers": 3,
#   "hits": 142,
#   "misses": 8,
#   "layer_stats": [
#     {"layer": 0, "backend": "memory", "hits": 130, "misses": 12},
#     {"layer": 1, "backend": "file",   "hits": 10,  "misses": 2},
#     {"layer": 2, "backend": "redis",  "hits": 2,   "misses": 0},
#   ]
# }
```

---

## 🔍 Backend Comparison

| Backend | Async | Persistent | Cross-Process | BlackDuck Safe | Install Extra |
|---|---|---|---|---|---|
| Memory | ✅ | ❌ | ❌ | ✅ | `[memory]` |
| File | ✅ | ✅ | ✅ | ✅ | none (stdlib only) |
| Redis | ✅ | ✅ | ✅ | ✅ | `[redis]` |
| PostgreSQL | ✅ | ✅ | ✅ | ✅ | `[postgres]` |
| Memcached | ✅ | ❌ | ✅ | ✅ | `[memcached]` |

---

## 🛡️ File Cache — Production Notes

`FileBackend` was built with four specific production concerns addressed:

**1. Any Python object (not just JSON)**
The default `PickleSerializer` supports `datetime`, `set`, custom classes — anything picklable. Switch to `JsonSerializer` only if you need human-readable files.
```python
# ✅ All of these work out of the box
await cache.set("ts", datetime.datetime.now())
await cache.set("tags", {"python", "caching", "async"})
await cache.set("obj", my_custom_object)
```

**2. Disk bloat prevention**
Files accumulate until TTL expires or you explicitly purge. Two options:
```python
# Option A: call manually (e.g. via a cron job or scheduler)
deleted = await cache.purge_expired()

# Option B: passive auto-purge every N seconds during writes
cache = FileBackend(FileConfig(directory="./cache", auto_purge_interval=3600))
```

**3. Windows concurrent write safety**
All writes use atomic `temp file → os.replace()` to avoid `PermissionError` under concurrent access on Windows.

**4. Windows path safety**
The cache directory is resolved to an absolute path at startup — no silent failures from deeply nested relative paths.

---

## 🚨 Error Handling

CacheStack wraps all backend errors in a clean exception hierarchy.

```python
from cachestack import CacheError, BackendUnavailableError

try:
    value = await cache.get("key")
except BackendUnavailableError:
    return default_value  # Redis/Postgres is down
except CacheError as e:
    logger.warning(f"Cache error: {e}")
```

Use `silent=True` on any backend config to suppress errors and return `None` instead of raising:

```python
# Never raises — returns None on any error
cache = FileBackend(FileConfig(directory="./cache", silent=True))
cache = RedisBackend(RedisConfig(dsn="redis://localhost", silent=True))
```

---

## 🔌 Writing a Custom Backend

Implement `BaseCache` and CacheStack will treat your backend like any built-in one — including full `TieredCache` support.

```python
from cachestack.base import BaseCache
from typing import Any, Dict, Optional

class MyCustomBackend(BaseCache):
    async def get(self, key: str) -> Optional[Any]: ...
    async def set(self, key: str, value: Any, ttl: Optional[int] = None) -> None: ...
    async def delete(self, key: str) -> None: ...
    async def exists(self, key: str) -> bool: ...
    async def clear(self) -> None: ...
    async def stats(self) -> Dict[str, Any]: ...

# Drop it into TieredCache like any other backend
cache = TieredCache([MemoryBackend(), MyCustomBackend()])
```

---

## 🧪 Running Tests

```bash
pip install pytest pytest-asyncio cachetools
pytest tests/ -v

# Expected: 36 passed
# (24 core tests + 12 file backend tests)
```

---

## 📤 Publishing to PyPI

```bash
pip install build twine

# Build the distribution
python -m build

# Upload to PyPI
twine upload dist/*
```

---

## 🤝 Contributing

Contributions are welcome!

1. Fork the repo on GitHub
2. Create a branch: `git checkout -b feature/my-feature`
3. Add tests for your change
4. Run `pytest tests/ -v` and ensure all 36 tests pass
5. Open a Pull Request with a clear description

---

## 📄 License

MIT — see [LICENSE](LICENSE).

---
