Metadata-Version: 2.2
Name: aiaio
Version: 0.0.8
Summary: aiaio
Author: abhishek
Project-URL: Homepage, https://github.com/abhishekkrthakur/aiaio
Project-URL: Issues, https://github.com/abhishekkrthakur/aiaio/issues
Keywords: aiaio
Classifier: Development Status :: 5 - Production/Stable
Classifier: License :: OSI Approved :: Apache Software License
Classifier: Operating System :: OS Independent
Classifier: Programming Language :: Python :: 3.10
Classifier: Programming Language :: Python :: 3.11
Classifier: Topic :: Scientific/Engineering :: Artificial Intelligence
Description-Content-Type: text/markdown
License-File: LICENSE
Requires-Dist: fastapi
Requires-Dist: uvicorn
Requires-Dist: loguru
Requires-Dist: jinja2
Requires-Dist: python-multipart
Requires-Dist: openai
Requires-Dist: websockets

# aiaio (AI-AI-O)

A lightweight, privacy-focused web UI for interacting with AI models. Supports both local and remote LLM deployments through OpenAI-compatible APIs.

![Screenshot](https://github.com/abhishekkrthakur/aiaio/blob/main/ui.png?raw=true)

## Features

- 🌓 Dark/Light mode support
- 💾 Local SQLite database for conversation storage
- 📁 File upload and processing (images, documents, etc.)
- ⚙️ Configurable model parameters through UI
- 🔒 Privacy-focused (all data stays local)
- 📱 Responsive design for mobile/desktop
- 🎨 Syntax highlighting for code blocks
- 📋 One-click code block copying
- 🔄 Real-time conversation updates
- 📝 Automatic conversation summarization
- 🎯 Customizable system prompts
- 🌐 WebSocket support for real-time updates
- 📦 Docker support for easy deploymen
- 📦 Multiple API endpoint support
- 📦 Multiple system prompt support

## Requirements


- Python 3.10+
- An OpenAI-compatible API endpoint (local or remote)

## Supported API Endpoints

aiaio works with any OpenAI-compatible API endpoint, including:

- OpenAI API
- vLLM
- Text Generation Inference (TGI)
- Hugging Face Inference Endpoints
- llama.cpp server
- LocalAI
- Custom OpenAI-compatible APIs

For example, you can serve llama 8b using vLLM using:

```bash
vllm serve Meta-Llama-3.1-8B-Instruct.Q4_K_M.gguf --tokenizer meta-llama/Llama-3.1-8B-Instruct --max_model_len 125000
```

and once the api is running, you can access it using aiaio ui.

## Installation

### Using pip

```bash
pip install aiaio
```

### From source

```bash
git clone https://github.com/abhishekkrthakur/aiaio.git
cd aiaio
pip install -e .
```

## Quick Start

1. Start the server:
```bash
aiaio app --host 127.0.0.1 --port 5000
```

2. Open your browser and navigate to `http://127.0.0.1:5000`

3. Configure your API endpoint and model settings in the UI

## Docker Usage

1. Build the Docker image:
```bash
docker build -t aiaio .
```

2. Run the container:
```bash
docker run --network host \
  -v /path/to/data:/data \
  aiaio
```

The `/data` volume mount is optional but recommended for persistent storage of the SQLite database and uploaded files.

## UI Configuration

All model and API settings can be configured through the UI:

### Model Parameters
- **Temperature** (0-2): Controls response randomness. Higher values make output more creative but less focused
- **Max Tokens** (1-32k): Maximum length of generated responses
- **Top P** (0-1): Controls diversity via nucleus sampling. Lower values make output more focused
- **Model Name**: Name/path of the model to use (depends on your API endpoint)

### API Configuration
- **Host**: URL of your OpenAI-compatible API endpoint
- **API Key**: Authentication key if required by your endpoint

These settings are stored in the local SQLite database and persist between sessions.

## File Handling

aiaio supports uploading and processing various file types, depending on the model's capabilities:

- Images (PNG, JPG, GIF, etc.)
- Documents (PDF, DOC, DOCX)
- Text files (TXT, CSV, JSON)
- Audio files (depends on model capabilities)
- Video files (depends on model capabilities)

Uploaded files are stored temporarily and can be referenced in conversations.

## Database Schema

aiaio uses SQLite for storage with the following main tables:

- `conversations`: Stores chat histories and summaries
- `messages`: Stores individual messages within conversations
- `attachments`: Stores file attachment metadata
- `settings`: Stores UI and model configuration

## Advanced Usage

### Custom System Prompts

Each conversation can have its own system prompt that guides the AI's behavior. Click the "System Prompt" section above the chat to customize it.

### Conversation Management

- Create new conversations using the "+ New Chat" button
- Switch between conversations in the left sidebar
- Delete conversations using the trash icon
- View conversation summaries in the sidebar

### Keyboard Shortcuts

- `Ctrl/Cmd + Enter`: Send message
- `Esc`: Clear input
- `Ctrl/Cmd + K`: Focus chat input
- `Ctrl/Cmd + /`: Toggle settings sidebar

## Development

```bash
# Clone the repository
git clone https://github.com/abhishekkrthakur/aiaio.git
cd aiaio

# Create a virtual environment
python -m venv venv
source venv/bin/activate  # or `venv\Scripts\activate` on Windows

# Install development dependencies
pip install -e ".[dev]"

# Run tests
pytest

# Run with auto-reload for development
uvicorn aiaio.app.app:app --reload --port 5000
```

## Contributing

Contributions are welcome! Please:

1. Fork the repository
2. Create a feature branch
3. Make your changes
4. Run the tests (`pytest`)
5. Submit a pull request

## License

Apache License 2.0 - see LICENSE file for details

## Acknowledgements

This project was primarily written with GitHub Copilot's assistance. While the human guided the development, Copilot generated much of the actual code.
