Metadata-Version: 2.4
Name: filechat
Version: 0.2.1
Summary: A local read-only AI coding assistant
Author-email: Milos Svana <milos@svana.name>
License-Expression: MIT
Requires-Python: >=3.12
Requires-Dist: einops>=0.8.1
Requires-Dist: faiss-cpu>=1.11.0.post1
Requires-Dist: mistralai>=1.9.3
Requires-Dist: onnxruntime>=1.22.1
Requires-Dist: pip-system-certs>=5.2
Requires-Dist: pydantic>=2.11.7
Requires-Dist: pyperclip>=1.9.0
Requires-Dist: textual>=5.3.0
Requires-Dist: tokenizers>=0.21.4
Requires-Dist: tqdm>=4.67.1
Requires-Dist: watchdog>=6.0.0
Provides-Extra: cuda
Requires-Dist: nvidia-cublas-cu12; extra == 'cuda'
Requires-Dist: nvidia-cuda-nvrtc-cu12; extra == 'cuda'
Requires-Dist: nvidia-cudnn-cu12; extra == 'cuda'
Requires-Dist: nvidia-cufft-cu12; extra == 'cuda'
Requires-Dist: nvidia-curand-cu12; extra == 'cuda'
Requires-Dist: onnxruntime-gpu; extra == 'cuda'
Provides-Extra: xpu
Requires-Dist: onnxruntime-openvino; extra == 'xpu'
Description-Content-Type: text/markdown

# FileChat

FileChat is an AI assistant designed to help users understand and improve their local projects.
It allows you to chat about files in your local folder while maintaining full control over your code.

Here is a short video:

https://github.com/user-attachments/assets/dd3c6617-b141-47ab-926e-c62abcc7b4a6

## Features

- **Project Indexing**: Creates a searchable index of your project files
- **Contextual Chat**: Ask questions about your project with AI that understands your codebase
- **Real-time Updates**: Automatically detects and indexes file changes
- **Configurable**: Customize which files to index and how to process them
- **Chat History**: ChatGPT-like chat history for each directory

## Installation

### Prerequisites

- Python 3.12 or higher
- A [Mistral AI](https://mistral.ai/) API key stored in the `MISTRAL_API_KEY` environment variable
- On Windows, you need [Visual C++ Redistributable](https://learn.microsoft.com/en-au/cpp/windows/latest-supported-vc-redist?view=msvc-170).
  It's very likely you have it already installed on your machine. 

### Option 1: Install from PyPI

You can use any Package management tool you like. Here is an example for `pip`:

```bash
pip install filechat
```

And here is an example of installing FileChat as a UV tool:

```bash
uv tool install filechat
```

**On Linux, you should also specify the hardware accelerator as an optional dependency**. We support `cpu`, `xpu` (Intel Arc), and `cuda`.
If you don't specify the accelerator, you'll get a version with CUDA support, which might be unnecessarily large. Here is an example of
installing FileChat with `xpu` support:

PIP:

```bash
pip install filechat[xpu]
```

UV Tool:

```bash
uv tool install filechat[xpu]
```

### Option 2: Clone the repository and use UV

1. Clone the repository:

```bash
git clone https://github.com/msvana/filechat
cd filechat
```

2. Install dependencies using [`uv`](https://docs.astral.sh/uv/):

```bash
uv sync
```

3. (Optional) Install GPU support if available:

```bash
# CUDA (NVIDIA)
uv sync --extra cuda

# XPU (Intel Arc)
uv sync --extra xpu
```

## Usage

```bash
filechat /path/to/your/project
```

## Configuration

On the first run, FileChat creates a configuration file at `~/.config/filechat.json`. Feel free to change it as you need.
Here is a full example:

```json
{
    "max_file_size_kb": 30,
    "ignored_dirs": [".git", "__pycache__", ".venv", ".pytest_cache", "node_modules"],
    "allowed_suffixes": [".txt", ".json", ".py", ".toml", ".html", ".md", ".js", ".ts", ".vue"],
    "index_store_path": "/home/milos/.cache/filechat",
    "model": "mistral-medium-2508",
    "api_key": "[MISTRAL_API_KEY]"
}
```

## Roadmap

### Short term (weeks)

- Add support for other models
- Add tools for browsing the filesystem
- Support CUDA on Windows

### Long term (months)

- Improve file retrieval (for example, via Graph RAG)
- Reimplement file indexing and querying in a compiled language
- Support important binary file types (images, PDFs)
- Add web search tools
