jeevesagent.vectorstore¶
Vector stores for semantic search over Chunk /
Document objects.
Unified async interface (modeled on LangChain’s VectorStore but
properly async-first and typed against our Chunk /
Document from jeevesagent.loader):
add()— embed + store chunks; returns their ids.delete()— remove by id.search()— top-k by cosine similarity + metadata filter.search_by_vector()— same with a precomputed query vector.
Implementations:
InMemoryVectorStore— default; zero-deps; cosine over a Python list. Great for dev / tests / small corpora.ChromaVectorStore— wrapschromadbfor persistent on-disk or hosted Chroma. Lazy import.PostgresVectorStore— wrapspgvectorviaasyncpg. Production durable. Lazy import.FAISSVectorStore— wrapsfaiss-cpufor fast in-memory ANN search over large corpora. Lazy import.
One-liner usage:
from jeevesagent import HashEmbedder
from jeevesagent.vectorstore import InMemoryVectorStore
from jeevesagent.loader import load, MarkdownChunker
vs = InMemoryVectorStore(embedder=HashEmbedder())
doc = load("research.pdf")
chunks = MarkdownChunker().split(doc.content, source=str(doc.metadata["source"]))
await vs.add(chunks)
results = await vs.search("what is RAG?", k=5)
for r in results:
print(f"{r.score:.3f}: {r.chunk.content[:100]}")
Optional dependencies:
pip install 'jeevesagent[vectorstore-chroma]'
pip install 'jeevesagent[vectorstore-postgres]'
pip install 'jeevesagent[vectorstore-faiss]'
pip install 'jeevesagent[vectorstore]' # all of the above
Submodules¶
Classes¶
Vector store backed by |
|
Vector store backed by |
|
In-process vector store backed by a Python list. |
|
Vector store backed by Postgres + |
|
One hit from |
|
Async protocol for vector stores. |
Package Contents¶
- class jeevesagent.vectorstore.ChromaVectorStore(embedder: jeevesagent.core.protocols.Embedder, *, collection_name: str = 'jeeves_vectors', persist_directory: str | None = None, client: Any = None)[source]¶
Vector store backed by
chromadb.- async add(chunks: list[jeevesagent.loader.base.Chunk], ids: list[str] | None = None) list[str][source]¶
- classmethod from_chunks(chunks: list[jeevesagent.loader.base.Chunk], *, embedder: jeevesagent.core.protocols.Embedder, ids: list[str] | None = None, collection_name: str = 'jeeves_vectors', persist_directory: str | None = None, client: Any = None) ChromaVectorStore[source]¶
- Async:
One-shot: construct a ChromaVectorStore + add
chunks.
- classmethod from_texts(texts: list[str], *, embedder: jeevesagent.core.protocols.Embedder, metadatas: list[dict[str, Any]] | None = None, ids: list[str] | None = None, collection_name: str = 'jeeves_vectors', persist_directory: str | None = None, client: Any = None) ChromaVectorStore[source]¶
- Async:
One-shot: construct a ChromaVectorStore from raw text strings (each becomes a
Chunkwith the matching metadata dict, or empty ifmetadatasis None).
- async search(query: str, *, k: int = 4, filter: collections.abc.Mapping[str, Any] | None = None, diversity: float | None = None) list[jeevesagent.vectorstore.base.SearchResult][source]¶
- async search_by_vector(vector: list[float], *, k: int = 4, filter: collections.abc.Mapping[str, Any] | None = None, diversity: float | None = None) list[jeevesagent.vectorstore.base.SearchResult][source]¶
- property embedder: jeevesagent.core.protocols.Embedder¶
- name = 'chroma'¶
- class jeevesagent.vectorstore.FAISSVectorStore(embedder: jeevesagent.core.protocols.Embedder, *, dimension: int | None = None, index_factory_string: str = 'HNSW32', metric: str = 'ip')[source]¶
Vector store backed by
faiss-cpu.- async add(chunks: list[jeevesagent.loader.base.Chunk], ids: list[str] | None = None) list[str][source]¶
- classmethod from_chunks(chunks: list[jeevesagent.loader.base.Chunk], *, embedder: jeevesagent.core.protocols.Embedder, ids: list[str] | None = None, dimension: int | None = None, index_factory_string: str = 'HNSW32', metric: str = 'ip') FAISSVectorStore[source]¶
- Async:
One-shot: construct a FAISSVectorStore + add
chunks.
- classmethod from_texts(texts: list[str], *, embedder: jeevesagent.core.protocols.Embedder, metadatas: list[dict[str, Any]] | None = None, ids: list[str] | None = None, dimension: int | None = None, index_factory_string: str = 'HNSW32', metric: str = 'ip') FAISSVectorStore[source]¶
- Async:
One-shot: construct a FAISSVectorStore from raw text strings (each becomes a
Chunkwith the matching metadata dict, or empty ifmetadatasis None).
- async search(query: str, *, k: int = 4, filter: collections.abc.Mapping[str, Any] | None = None, diversity: float | None = None) list[jeevesagent.vectorstore.base.SearchResult][source]¶
- async search_by_vector(vector: list[float], *, k: int = 4, filter: collections.abc.Mapping[str, Any] | None = None, diversity: float | None = None) list[jeevesagent.vectorstore.base.SearchResult][source]¶
- property embedder: jeevesagent.core.protocols.Embedder¶
- name = 'faiss'¶
- class jeevesagent.vectorstore.InMemoryVectorStore(embedder: jeevesagent.core.protocols.Embedder)[source]¶
In-process vector store backed by a Python list.
- async add(chunks: list[jeevesagent.loader.base.Chunk], ids: list[str] | None = None) list[str][source]¶
- classmethod from_chunks(chunks: list[jeevesagent.loader.base.Chunk], *, embedder: jeevesagent.core.protocols.Embedder, ids: list[str] | None = None) InMemoryVectorStore[source]¶
- Async:
One-shot: construct an InMemoryVectorStore + add
chunks.
- classmethod from_texts(texts: list[str], *, embedder: jeevesagent.core.protocols.Embedder, metadatas: list[dict[str, Any]] | None = None, ids: list[str] | None = None) InMemoryVectorStore[source]¶
- Async:
One-shot: construct an InMemoryVectorStore from raw text strings (each becomes a
Chunkwith the matching metadata dict, or empty ifmetadatasis None).
- classmethod load(path: str | pathlib.Path, *, embedder: jeevesagent.core.protocols.Embedder) InMemoryVectorStore[source]¶
- Async:
Restore a store previously
save()-d. Pass the same embedder kind/dimensions or queries will produce nonsense scores.
- async save(path: str | pathlib.Path) None[source]¶
Write the full store (chunks + vectors + ids) to a JSON file. The embedder is NOT serialized — supply the same embedder when calling
load().
- async search(query: str, *, k: int = 4, filter: collections.abc.Mapping[str, Any] | None = None, diversity: float | None = None) list[jeevesagent.vectorstore.base.SearchResult][source]¶
- async search_by_vector(vector: list[float], *, k: int = 4, filter: collections.abc.Mapping[str, Any] | None = None, diversity: float | None = None) list[jeevesagent.vectorstore.base.SearchResult][source]¶
- async search_hybrid(query: str, *, k: int = 4, filter: collections.abc.Mapping[str, Any] | None = None, alpha: float = 0.5) list[jeevesagent.vectorstore.base.SearchResult][source]¶
Hybrid lexical (BM25) + vector search via RRF.
alphais in [0, 1]: 0 = pure BM25, 1 = pure vector, 0.5 = even weighting (RRF default). Both rankings are computed independently and fused by Reciprocal Rank Fusion, then the top-ksurvivors are returned.Embeddings catch semantic similarity (“automobile” ↔ “car”), BM25 catches exact-term hits (model names, error codes, person names) — together they outperform either alone on most retrieval benchmarks.
- property embedder: jeevesagent.core.protocols.Embedder¶
- name = 'in-memory'¶
- class jeevesagent.vectorstore.PostgresVectorStore(embedder: jeevesagent.core.protocols.Embedder, *, dsn: str, table: str = 'jeeves_vectors', dimension: int | None = None)[source]¶
Vector store backed by Postgres +
pgvector.- async add(chunks: list[jeevesagent.loader.base.Chunk], ids: list[str] | None = None) list[str][source]¶
- classmethod from_chunks(chunks: list[jeevesagent.loader.base.Chunk], *, embedder: jeevesagent.core.protocols.Embedder, ids: list[str] | None = None, dsn: str, table: str = 'jeeves_vectors', dimension: int | None = None) PostgresVectorStore[source]¶
- Async:
One-shot: construct a PostgresVectorStore + add
chunks.
- classmethod from_texts(texts: list[str], *, embedder: jeevesagent.core.protocols.Embedder, metadatas: list[dict[str, Any]] | None = None, ids: list[str] | None = None, dsn: str, table: str = 'jeeves_vectors', dimension: int | None = None) PostgresVectorStore[source]¶
- Async:
One-shot: construct a PostgresVectorStore from raw text strings (each becomes a
Chunkwith the matching metadata dict, or empty ifmetadatasis None).
- async search(query: str, *, k: int = 4, filter: collections.abc.Mapping[str, Any] | None = None, diversity: float | None = None) list[jeevesagent.vectorstore.base.SearchResult][source]¶
- async search_by_vector(vector: list[float], *, k: int = 4, filter: collections.abc.Mapping[str, Any] | None = None, diversity: float | None = None) list[jeevesagent.vectorstore.base.SearchResult][source]¶
- property embedder: jeevesagent.core.protocols.Embedder¶
- name = 'postgres'¶
- class jeevesagent.vectorstore.SearchResult[source]¶
One hit from
VectorStore.search().chunk— the matched chunk (with its full metadata).score— similarity in [-1, 1] for cosine; backend- specific for other distance metrics. Higher = more similar.id— the store-assigned id (so callers candelete()orget_by_ids()later).
- class jeevesagent.vectorstore.VectorStore[source]¶
Bases:
ProtocolAsync protocol for vector stores.
Six methods cover the lifecycle: add (embed + store), delete, search (by query string), search_by_vector (precomputed), count, get_by_ids.
Backends that aren’t natively async (FAISS, Chroma) wrap their sync calls in
anyio.to_thread.run_sync()so they don’t block the event loop.- async add(chunks: list[jeevesagent.loader.base.Chunk], ids: list[str] | None = None) list[str][source]¶
Embed + store
chunks. Returns the assigned ids (caller-provided or generated).
- async delete(ids: list[str]) None[source]¶
Remove the named chunks. Unknown ids are silently skipped (idempotent).
- async get_by_ids(ids: list[str]) list[jeevesagent.loader.base.Chunk][source]¶
Fetch chunks by id, in the same order as
ids. Unknown ids are skipped (the result may be shorter than the input).
- async search(query: str, *, k: int = 4, filter: collections.abc.Mapping[str, Any] | None = None, diversity: float | None = None) list[SearchResult][source]¶
Embed
queryand return the top-kchunks ranked by similarity.filter(optional) restricts candidates by metadata.diversity(optional, 0..1) enables MMR reranking for varied results.