add openai compatible endpoint for chat completions

This commit is contained in:
geoffsee
2025-06-05 20:31:59 -04:00
committed by Geoff Seemueller
parent 3b4c8b045a
commit 8a3c0797c3
6 changed files with 965 additions and 78 deletions

View File

@@ -1,10 +1,27 @@
# open-web-agent-rs
A Rust-based web agent with local inference capabilities.
## Components
### Local Inference Engine
The [Local Inference Engine](./local_inference_engine/README.md) provides a way to run large language models locally. It supports both CLI mode for direct text generation and server mode with an OpenAI-compatible API.
Features:
- Run Gemma models locally (1B, 2B, 7B, 9B variants)
- CLI mode for direct text generation
- Server mode with OpenAI-compatible API
- Support for various model configurations (base, instruction-tuned)
- Metal acceleration on macOS
See the [Local Inference Engine README](./local_inference_engine/README.md) for detailed usage instructions.
### Web Server
Server is being converted to MCP. Things are probably broken.
```text
bun i
bun dev
```