Metadata-Version: 2.4
Name: swisper-studio-sdk
Version: 0.6.1
Summary: SwisperStudio SDK - Zero-config tracing for LangGraph applications
Author: Fintama SwisperStudio Team
License: Proprietary - Fintama Internal Use Only
Project-URL: Homepage, https://github.com/Fintama/swisper_studio
Project-URL: Repository, https://github.com/Fintama/swisper_studio
Project-URL: Documentation, https://github.com/Fintama/swisper_studio/tree/main/sdk
Requires-Python: >=3.11
Description-Content-Type: text/markdown
Requires-Dist: httpx>=0.25.2
Requires-Dist: langgraph<2.0.0,>=1.0.0
Requires-Dist: langgraph-checkpoint>=2.1.0
Requires-Dist: redis>=5.0.0
Provides-Extra: dev
Requires-Dist: pytest>=7.4.3; extra == "dev"
Requires-Dist: pytest-asyncio>=0.21.1; extra == "dev"

# SwisperStudio SDK

Simple, high-performance integration for tracing Swisper LangGraph applications.

**v0.5.4 - Full LLM Telemetry:**
- 🚀 **50x faster** - 500ms → 10ms overhead (Redis Streams)
- 🧠 **LLM reasoning** - See thinking process (`<think>...</think>`)
- 📊 **Full LLM telemetry** - Prompts, responses, token usage
- 📡 **Connection status** - Heartbeat-based health monitoring
- ⚙️ **Per-node config** - Fine-grained control

## Installation

### From PyPI (Recommended)

```bash
pip install swisper-studio-sdk>=0.5.4
```

That's it! No authentication needed.

### From Source (Development)

```bash
git clone https://github.com/Fintama/swisper_studio.git
cd swisper_studio/sdk
pip install -e .
```

**Note:** Source installation requires Fintama GitHub organization access.

## Quick Start (30 seconds)

### 1. Initialize at Startup (Redis Streams)

```python
# In your main.py or startup code
from swisper_studio_sdk import safe_initialize, wrap_llm_adapter

# Async initialization (in lifespan or startup) - RECOMMENDED
status = await safe_initialize(
    redis_url="redis://swisper_studio_redis:6379",  # SwisperStudio Redis
    project_id="your-project-id",                    # From SwisperStudio
    enabled=True
)

if status["initialized"]:
    logger.info("SwisperStudio tracing enabled")
    
    # CRITICAL: Wrap LLM adapters to capture prompts/responses/tokens
    wrap_llm_adapter()
else:
    logger.info("Tracing disabled (Swisper continues normally)")
```

**Note:** `safe_initialize()` NEVER blocks or raises exceptions. If Redis is unavailable, Swisper continues normally with tracing disabled.

**⚠️ IMPORTANT:** You MUST call `wrap_llm_adapter()` after successful initialization to capture LLM telemetry (prompts, responses, token usage). Without this, LLM-calling nodes will show as SPAN instead of GENERATION.

### 2. ONE LINE CHANGE to Enable Graph Tracing

```python
# Before:
from langgraph.graph import StateGraph
graph = StateGraph(GlobalSupervisorState)

# After (change ONE line):
from swisper_studio_sdk import create_traced_graph
graph = create_traced_graph(GlobalSupervisorState, trace_name="supervisor")

# All nodes added to this graph are automatically traced!
```

### 3. Add Nodes as Normal

```python
# Add nodes - they're automatically traced!
graph.add_node("intent_classification", intent_classification_node)
graph.add_node("memory", memory_node)
graph.add_node("planner", planner_node)
graph.add_node("ui_node", ui_node)

# Compile and run as usual
app = graph.compile()
result = await app.ainvoke(initial_state)

# All executions are now traced to SwisperStudio! 🎉
```

## Two-Part Setup Explained

SwisperStudio tracing requires TWO components:

| Component | Purpose | What it captures |
|-----------|---------|------------------|
| `create_traced_graph()` | Wraps graph nodes | Execution flow, state, timing |
| `wrap_llm_adapter()` | Wraps LLM adapters | Prompts, responses, tokens |

**Both are required for full observability:**

```python
# ✅ CORRECT - Full observability
await safe_initialize(...)
wrap_llm_adapter()  # Captures LLM telemetry
graph = create_traced_graph(...)  # Captures node execution

# ❌ WRONG - Missing LLM telemetry
await safe_initialize(...)
# wrap_llm_adapter() NOT called!
graph = create_traced_graph(...)  # Nodes show as SPAN, not GENERATION
```

## Features

### **Core Features:**
- ✅ **One-line integration** - `create_traced_graph()` instead of `StateGraph()`
- ✅ **Auto-instrumentation** - All nodes automatically traced
- ✅ **State capture** - Captures input/output state at each node
- ✅ **Error tracking** - Captures exceptions and error messages
- ✅ **Nested observations** - Supports parent-child relationships
- ✅ **Zero boilerplate** - No decorators needed on individual nodes

### **v0.5.x Features:**
- ✅ **Redis Streams** - 50x faster than HTTP (500ms → 10ms)
- ✅ **LLM Reasoning** - Captures `<think>...</think>` tags from DeepSeek R1, o1, etc.
- ✅ **Streaming Support** - Captures full responses from streaming LLM calls
- ✅ **Connection Status** - Verifies SwisperStudio consumer is running
- ✅ **Per-Node Config** - Enable/disable reasoning per node
- ✅ **Memory Safety** - Auto-cleanup prevents memory leaks
- ✅ **Full LLM Telemetry** (v0.5.4) - Wraps LLMAdapterFactory directly for reliable capture

## Advanced Usage

### LLM Reasoning Capture

Control reasoning capture per node:

```python
from swisper_studio_sdk import traced

# Enable reasoning with custom length limit
@traced("classify_intent", capture_reasoning=True, reasoning_max_length=20000)
async def classify_intent_node(state):
    # Captures <think>...</think> tags (up to 20KB)
    return state

# Disable reasoning for specific nodes
@traced("memory_node", capture_reasoning=False)
async def memory_node(state):
    # No reasoning captured (faster, less data)
    return state

# Use defaults (reasoning enabled, 50KB limit)
@traced("global_planner")
async def global_planner_node(state):
    return state
```

**What gets captured:**
- ✅ LLM prompts (system + user messages)
- ✅ Reasoning process (`<think>...</think>` tags)
- ✅ Final responses (structured output or streaming)
- ✅ Token usage (prompt + completion)

**Supported models:**
- DeepSeek R1 (with reasoning)
- OpenAI o1/o3 (with reasoning)
- GPT-4, Claude, Llama (no reasoning, just prompts + responses)

### Manual Tracing (Optional)

For fine-grained control, use `@traced` decorator:

```python
from swisper_studio_sdk import traced

# Full control over observation
@traced(
    name="intent_classification",
    observation_type="GENERATION",
    capture_reasoning=True,
    reasoning_max_length=10000
)
async def intent_classification_node(state):
    return state
```

### Observation Types

- `AUTO` - Auto-detect based on LLM data (default, recommended)
- `SPAN` - Generic execution span
- `GENERATION` - LLM generation
- `EVENT` - Point-in-time event
- `TOOL` - Tool call
- `AGENT` - Agent execution

## Architecture

### Redis Streams (v0.4.0)

```
Your App (Swisper)         Redis Stream              SwisperStudio
       │                        │                           │
  @traced decorator             │                           │
       │                        │                           │
  XADD event (1-2ms) ──────────→ │                           │
       │                        │                           │
  Return immediately            │                           │
  (zero latency!)               │                           │
                                │   Consumer reads batch    │
                                │ ←─────────────────────────┤
                                │                           │
                                │   Store in PostgreSQL     │
                                │ ──────────────────────────→
```

**Benefits:**
- **50x faster** than HTTP (500ms → 10ms overhead)
- **No race conditions** (ordered stream delivery)
- **Reliable** (persistent queue, automatic retry)
- **Scalable** (100k+ events/sec)

### How It Works

1. `create_traced_graph()` monkey-patches `add_node()` to auto-wrap functions
2. `@traced` decorator publishes events to Redis Streams (1-2ms)
3. SwisperStudio consumer reads from stream and stores in database
4. Zero user-facing latency (fire-and-forget pattern)

## Configuration

### Required Settings

```python
# In your config.py or .env
SWISPER_STUDIO_REDIS_URL: str = "redis://redis:6379"
SWISPER_STUDIO_PROJECT_ID: str = "your-project-id"
SWISPER_STUDIO_STREAM_NAME: str = "observability:events"
```

### Optional Settings

```python
# Reasoning capture
SWISPER_STUDIO_CAPTURE_REASONING: bool = True
SWISPER_STUDIO_REASONING_MAX_LENGTH: int = 50000  # 50 KB

# Connection verification
SWISPER_STUDIO_VERIFY_CONSUMER: bool = True  # Check consumer health
```

## Requirements

- Python 3.11+
- LangGraph >= 1.0.0, < 2.0.0
- langgraph-checkpoint >= 2.1.0 (v0.5.2 removed upper bound for HITL compatibility)
- httpx >= 0.25.2
- redis >= 5.0.0

## Migration

### From v0.5.x to v0.5.4
No code changes required. Just update the SDK:
```bash
pip install swisper-studio-sdk==0.5.4
```

### From v0.4.x or earlier
1. Update SDK: `pip install swisper-studio-sdk>=0.5.4`
2. Add `wrap_llm_adapter()` after `safe_initialize()`:

```python
status = await safe_initialize(...)
if status["initialized"]:
    wrap_llm_adapter()  # ADD THIS LINE
```

### From v0.3.x
See [SDK_MIGRATION_v0.3.4_to_v0.4.0.md](../../SDK_MIGRATION_v0.3.4_to_v0.4.0.md)

**Migration time:** ~5 minutes  
**Breaking changes:** None (backward compatible)

## Troubleshooting

### LLM nodes showing as SPAN instead of GENERATION

**Symptom:** LLM-calling nodes (classify_intent, global_planner, etc.) show as SPAN with no token data.

**Cause:** `wrap_llm_adapter()` was not called after initialization.

**Fix:**
```python
status = await safe_initialize(...)
if status["initialized"]:
    wrap_llm_adapter()  # Must be called!
```

**Verify in logs:**
```
✅ LLMAdapterFactory wrapped for LLM telemetry capture
✅ LLM adapter wrapped for prompt capture (2 adapter(s))
```

### Traces not appearing in SwisperStudio

1. Check Redis connectivity in initialization logs
2. Verify project_id matches your SwisperStudio project
3. Ensure SwisperStudio backend consumer is running

### Token counts showing as 0

This is normal for streaming responses where the LLM doesn't return usage data.
Use `get_structured_output()` for accurate token counts.

## License

MIT

