mirror of
https://github.com/geoffsee/predict-otron-9001.git
synced 2025-09-08 22:46:44 +00:00
41 lines
1.2 KiB
Markdown
41 lines
1.2 KiB
Markdown
# chat-ui
|
|
|
|
A WASM-based web chat interface for the predict-otron-9000 AI platform.
|
|
|
|
## Overview
|
|
|
|
The chat-ui provides a real-time web interface for interacting with language models through the predict-otron-9000 server. Built with Leptos and compiled to WebAssembly, it offers a modern chat experience with streaming response support.
|
|
|
|
## Features
|
|
|
|
- Real-time chat interface with the inference server
|
|
- Streaming response support
|
|
- Conversation history
|
|
- Responsive web design
|
|
- WebAssembly-powered for optimal performance
|
|
|
|
## Building and Running
|
|
|
|
### Prerequisites
|
|
- Rust toolchain with WASM target: `rustup target add wasm32-unknown-unknown`
|
|
- The predict-otron-9000 server must be running on port 8080
|
|
|
|
### Development Server
|
|
```bash
|
|
cd crates/chat-ui
|
|
./run.sh
|
|
```
|
|
|
|
This starts the development server on port 8788 with auto-reload capabilities.
|
|
|
|
### Usage
|
|
1. Start the predict-otron-9000 server: `./scripts/run.sh`
|
|
2. Start the chat-ui: `cd crates/chat-ui && ./run.sh`
|
|
3. Navigate to `http://localhost:8788`
|
|
4. Start chatting with your AI models!
|
|
|
|
## Technical Details
|
|
- Built with Leptos framework
|
|
- Compiled to WebAssembly for browser execution
|
|
- Communicates with predict-otron-9000 API via HTTP
|
|
- Sets required RUSTFLAGS for WebAssembly getrandom support |