mirror of
https://github.com/geoffsee/open-gsio.git
synced 2025-09-08 22:56:46 +00:00
run format
This commit is contained in:

committed by
Geoff Seemueller

parent
02c3253343
commit
f76301d620
58
LEGACY.md
58
LEGACY.md
@@ -1,60 +1,60 @@
|
||||
Legacy Development History
|
||||
---
|
||||
## Legacy Development History
|
||||
|
||||
The source code of open-gsio was drawn from the source code of my personal website. That commit history was contaminated early on with secrets. `open-gsio` is a refinement of those sources. A total of 367 commits were submitted to the main branch of the upstream source repository between August 2024 and May 2025.
|
||||
|
||||
#### **May 2025**
|
||||
|
||||
* Added **seemueller.ai** link to UI sidebar.
|
||||
* Global config/markdown guide clean‑up; patched a critical forgotten bug.
|
||||
- Added **seemueller.ai** link to UI sidebar.
|
||||
- Global config/markdown guide clean‑up; patched a critical forgotten bug.
|
||||
|
||||
#### **Apr 2025**
|
||||
|
||||
* **CI/CD overhaul**: auto‑deploy to dev & staging, Bun adoption as package manager, streamlined block‑list workflow (now auto‑updates via VPN blocker).
|
||||
* New 404 error page; multiple robots.txt and editor‑resize fixes; removed dead/duplicate code.
|
||||
- **CI/CD overhaul**: auto‑deploy to dev & staging, Bun adoption as package manager, streamlined block‑list workflow (now auto‑updates via VPN blocker).
|
||||
- New 404 error page; multiple robots.txt and editor‑resize fixes; removed dead/duplicate code.
|
||||
|
||||
#### **Mar 2025**
|
||||
|
||||
* Introduced **model‑specific `max_tokens`** handling and plugged in **Cloudflare AI models** for testing.
|
||||
* Bundle size minimised (re‑enabled minifier, smaller vendor set).
|
||||
- Introduced **model‑specific `max_tokens`** handling and plugged in **Cloudflare AI models** for testing.
|
||||
- Bundle size minimised (re‑enabled minifier, smaller vendor set).
|
||||
|
||||
#### **Feb 2025**
|
||||
|
||||
* **Full theme system** (runtime switching, Centauri theme, server‑saved prefs).
|
||||
* Tightened MobX typing for messages; responsive break‑points & input scaling repaired.
|
||||
* Dropped legacy document API; general folder restructure.
|
||||
- **Full theme system** (runtime switching, Centauri theme, server‑saved prefs).
|
||||
- Tightened MobX typing for messages; responsive break‑points & input scaling repaired.
|
||||
- Dropped legacy document API; general folder restructure.
|
||||
|
||||
#### **Jan 2025**
|
||||
|
||||
* **Rate‑limit middleware**, larger KV/R2 storage quota.
|
||||
* Switched default model → *llama‑v3p1‑70b‑instruct*; pluggable model handlers.
|
||||
* Added **KaTeX fonts** & **Marked.js** for rich math/markdown.
|
||||
* Fireworks key rotation; deprecated Google models removed.
|
||||
- **Rate‑limit middleware**, larger KV/R2 storage quota.
|
||||
- Switched default model → _llama‑v3p1‑70b‑instruct_; pluggable model handlers.
|
||||
- Added **KaTeX fonts** & **Marked.js** for rich math/markdown.
|
||||
- Fireworks key rotation; deprecated Google models removed.
|
||||
|
||||
#### **Dec 2024**
|
||||
|
||||
* Major package upgrades; **CodeHighlighter** now supports HTML/JSX/TS(X)/Zig.
|
||||
* Refactored streaming + markdown renderer; Android‑specific padding fixes.
|
||||
* Reset default chat model to **gpt‑4o**; welcome message & richer search‑intent logic.
|
||||
- Major package upgrades; **CodeHighlighter** now supports HTML/JSX/TS(X)/Zig.
|
||||
- Refactored streaming + markdown renderer; Android‑specific padding fixes.
|
||||
- Reset default chat model to **gpt‑4o**; welcome message & richer search‑intent logic.
|
||||
|
||||
#### **Nov 2024**
|
||||
|
||||
* **Fireworks API** + agent server; first‑class support for **Anthropic** & **GROQ** models (incl. attachments).
|
||||
* **VPN blocker** shipped with CIDR validation and dedicated GitHub Action.
|
||||
* Live search buffering, feedback modal, smarter context preprocessing.
|
||||
- **Fireworks API** + agent server; first‑class support for **Anthropic** & **GROQ** models (incl. attachments).
|
||||
- **VPN blocker** shipped with CIDR validation and dedicated GitHub Action.
|
||||
- Live search buffering, feedback modal, smarter context preprocessing.
|
||||
|
||||
#### **Oct 2024**
|
||||
|
||||
* Rolled out **image generation** + picker for image models.
|
||||
* Deployed **ETH payment processor** & deposit‑address flow.
|
||||
* Introduced few‑shot prompting library; analytics worker refactor; Halloween prompt.
|
||||
* Extensive mobile‑UX polish and bundling/worker config updates.
|
||||
- Rolled out **image generation** + picker for image models.
|
||||
- Deployed **ETH payment processor** & deposit‑address flow.
|
||||
- Introduced few‑shot prompting library; analytics worker refactor; Halloween prompt.
|
||||
- Extensive mobile‑UX polish and bundling/worker config updates.
|
||||
|
||||
#### **Sep 2024**
|
||||
|
||||
* End‑to‑end **math rendering** (KaTeX) and **GitHub‑flavoured markdown**.
|
||||
* Migrated chat state to **MobX**; launched analytics service & metrics worker.
|
||||
* Switched build minifier to **esbuild**; tokenizer limits enforced; gradient sidebar & cookie‑consent manager added.
|
||||
- End‑to‑end **math rendering** (KaTeX) and **GitHub‑flavoured markdown**.
|
||||
- Migrated chat state to **MobX**; launched analytics service & metrics worker.
|
||||
- Switched build minifier to **esbuild**; tokenizer limits enforced; gradient sidebar & cookie‑consent manager added.
|
||||
|
||||
#### **Aug 2024**
|
||||
|
||||
* **Initial MVP**: iMessage‑style chat UI, websocket prototype, Google Analytics, Cloudflare bindings, base worker‑site scaffold.
|
||||
- **Initial MVP**: iMessage‑style chat UI, websocket prototype, Google Analytics, Cloudflare bindings, base worker‑site scaffold.
|
||||
|
76
README.md
76
README.md
@@ -1,12 +1,13 @@
|
||||
# open-gsio
|
||||
|
||||
[](https://github.com/geoffsee/open-gsio/actions/workflows/test.yml)
|
||||
[](https://opensource.org/licenses/MIT)
|
||||
</br>
|
||||
|
||||
<p align="center">
|
||||
<img src="https://github.com/user-attachments/assets/620d2517-e7be-4bb0-b2b7-3aa0cba37ef0" width="250" />
|
||||
</p>
|
||||
|
||||
|
||||
This is a full-stack Conversational AI.
|
||||
|
||||
## Table of Contents
|
||||
@@ -23,7 +24,6 @@ This is a full-stack Conversational AI.
|
||||
- [Acknowledgments](#acknowledgments)
|
||||
- [License](#license)
|
||||
|
||||
|
||||
## Installation
|
||||
|
||||
1. `bun i && bun test:all`
|
||||
@@ -33,6 +33,7 @@ This is a full-stack Conversational AI.
|
||||
> Note: it should be possible to use pnpm in place of bun.
|
||||
|
||||
## Deployment
|
||||
|
||||
1. Setup KV_STORAGE binding in `packages/server/wrangler.jsonc`
|
||||
1. [Add keys in secrets.json](https://console.groq.com/keys)
|
||||
1. Run `bun run deploy && bun run deploy:secrets && bun run deploy`
|
||||
@@ -40,11 +41,14 @@ This is a full-stack Conversational AI.
|
||||
> Note: Subsequent deployments should omit `bun run deploy:secrets`
|
||||
|
||||
## Local Inference
|
||||
|
||||
> Local inference is supported for Ollama and mlx-omni-server. OpenAI compatible servers can be used by overriding OPENAI_API_KEY and OPENAI_API_ENDPOINT.
|
||||
|
||||
### mlx-omni-server
|
||||
|
||||
(default) (Apple Silicon Only)
|
||||
~~~bash
|
||||
|
||||
```bash
|
||||
# (prereq) install mlx-omni-server
|
||||
brew tap seemueller-io/tap
|
||||
brew install seemueller-io/tap/mlx-omni-server
|
||||
@@ -52,10 +56,11 @@ brew install seemueller-io/tap/mlx-omni-server
|
||||
bun run openai:local mlx-omni-server # Start mlx-omni-server
|
||||
bun run openai:local:configure # Configure connection
|
||||
bun run server:dev # Restart server
|
||||
~~~
|
||||
```
|
||||
|
||||
#### Adding models for local inference (Apple Silicon)
|
||||
|
||||
~~~bash
|
||||
```bash
|
||||
# ensure mlx-omni-server is running
|
||||
|
||||
# See https://huggingface.co/mlx-community for available models
|
||||
@@ -67,21 +72,22 @@ curl http://localhost:10240/v1/chat/completions \
|
||||
\"model\": \"$MODEL_TO_ADD\",
|
||||
\"messages\": [{\"role\": \"user\", \"content\": \"Hello\"}]
|
||||
}"
|
||||
~~~
|
||||
```
|
||||
|
||||
### Ollama
|
||||
~~~bash
|
||||
|
||||
```bash
|
||||
bun run openai:local ollama # Start ollama server
|
||||
bun run openai:local:configure # Configure connection
|
||||
bun run server:dev # Restart server
|
||||
~~~
|
||||
```
|
||||
|
||||
#### Adding models for local inference (ollama)
|
||||
|
||||
~~~bash
|
||||
```bash
|
||||
# See https://ollama.com/library for available models
|
||||
use the ollama web ui @ http://localhost:8080
|
||||
~~~
|
||||
|
||||
```
|
||||
|
||||
## Testing
|
||||
|
||||
@@ -89,44 +95,44 @@ Tests are located in `__tests__` directories next to the code they test. Testing
|
||||
|
||||
> `bun test:all` will run all tests
|
||||
|
||||
|
||||
## Troubleshooting
|
||||
|
||||
1. `bun clean`
|
||||
1. `bun i`
|
||||
1. `bun server:dev`
|
||||
1. `bun client:dev`
|
||||
1. Submit an issue
|
||||
|
||||
History
|
||||
---
|
||||
## History
|
||||
|
||||
A high-level overview for the development history of the parent repository, [geoff-seemueller-io](https://geoff.seemueller.io), is provided in [LEGACY.md](./LEGACY.md).
|
||||
|
||||
## Acknowledgments
|
||||
|
||||
I would like to express gratitude to the following projects, libraries, and individuals that have contributed to making open-gsio possible:
|
||||
|
||||
- [TypeScript](https://www.typescriptlang.org/) - Primary programming language
|
||||
- [React](https://react.dev/) - UI library for building the frontend
|
||||
- [Vike](https://vike.dev/) - Framework for server-side rendering and routing
|
||||
- [Cloudflare Workers](https://developers.cloudflare.com/workers/) - Serverless execution environment
|
||||
- [Bun](https://bun.sh/) - JavaScript runtime and toolkit
|
||||
- [itty-router](https://github.com/kwhitley/itty-router) - Lightweight router for serverless environments
|
||||
- [MobX-State-Tree](https://mobx-state-tree.js.org/) - State management solution
|
||||
- [OpenAI SDK](https://github.com/openai/openai-node) - Client for AI model integration
|
||||
- [Vitest](https://vitest.dev/) - Testing framework
|
||||
- [OpenAI](https://github.com/openai)
|
||||
- [Groq](https://console.groq.com/) - Fast inference API
|
||||
- [Anthropic](https://www.anthropic.com/) - Creator of Claude models
|
||||
- [Fireworks](https://fireworks.ai/) - AI inference platform
|
||||
- [XAI](https://x.ai/) - Creator of Grok models
|
||||
- [Cerebras](https://www.cerebras.net/) - AI compute and models
|
||||
- [(madroidmaq) MLX Omni Server](https://github.com/madroidmaq/mlx-omni-server) - Open-source high-performance inference for Apple Silicon
|
||||
- [MLX](https://github.com/ml-explore/mlx) - An array framework for Apple silicon
|
||||
- [Ollama](https://github.com/ollama/ollama) - Versatile solution for self-hosting models
|
||||
|
||||
- [TypeScript](https://www.typescriptlang.org/) - Primary programming language
|
||||
- [React](https://react.dev/) - UI library for building the frontend
|
||||
- [Vike](https://vike.dev/) - Framework for server-side rendering and routing
|
||||
- [Cloudflare Workers](https://developers.cloudflare.com/workers/) - Serverless execution environment
|
||||
- [Bun](https://bun.sh/) - JavaScript runtime and toolkit
|
||||
- [itty-router](https://github.com/kwhitley/itty-router) - Lightweight router for serverless environments
|
||||
- [MobX-State-Tree](https://mobx-state-tree.js.org/) - State management solution
|
||||
- [OpenAI SDK](https://github.com/openai/openai-node) - Client for AI model integration
|
||||
- [Vitest](https://vitest.dev/) - Testing framework
|
||||
- [OpenAI](https://github.com/openai)
|
||||
- [Groq](https://console.groq.com/) - Fast inference API
|
||||
- [Anthropic](https://www.anthropic.com/) - Creator of Claude models
|
||||
- [Fireworks](https://fireworks.ai/) - AI inference platform
|
||||
- [XAI](https://x.ai/) - Creator of Grok models
|
||||
- [Cerebras](https://www.cerebras.net/) - AI compute and models
|
||||
- [(madroidmaq) MLX Omni Server](https://github.com/madroidmaq/mlx-omni-server) - Open-source high-performance inference for Apple Silicon
|
||||
- [MLX](https://github.com/ml-explore/mlx) - An array framework for Apple silicon
|
||||
- [Ollama](https://github.com/ollama/ollama) - Versatile solution for self-hosting models
|
||||
|
||||
## License
|
||||
~~~text
|
||||
|
||||
```text
|
||||
MIT License
|
||||
|
||||
Copyright (c) 2025 Geoff Seemueller
|
||||
@@ -148,4 +154,4 @@ AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
|
||||
LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
|
||||
OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
|
||||
SOFTWARE.
|
||||
~~~
|
||||
```
|
||||
|
@@ -38,5 +38,6 @@
|
||||
},
|
||||
"peerDependencies": {
|
||||
"typescript": "^5"
|
||||
}
|
||||
},
|
||||
"packageManager": "pnpm@10.10.0+sha512.d615db246fe70f25dcfea6d8d73dee782ce23e2245e3c4f6f888249fb568149318637dca73c2c5c8ef2a4ca0d5657fb9567188bfab47f566d1ee6ce987815c39"
|
||||
}
|
||||
|
@@ -4,10 +4,6 @@
|
||||
"outDir": "dist",
|
||||
"rootDir": "."
|
||||
},
|
||||
"include": [
|
||||
"*.ts"
|
||||
],
|
||||
"exclude": [
|
||||
"node_modules"
|
||||
]
|
||||
"include": ["*.ts"],
|
||||
"exclude": ["node_modules"]
|
||||
}
|
43
packages/client/public/cfga.min.js
vendored
43
packages/client/public/cfga.min.js
vendored
@@ -15,30 +15,29 @@
|
||||
};
|
||||
function s() {
|
||||
var i = [
|
||||
g(m(4)) + "=" + g(m(6)),
|
||||
"ga=" + t.ga_tid,
|
||||
"dt=" + r(e.title),
|
||||
"de=" + r(e.characterSet || e.charset),
|
||||
"dr=" + r(e.referrer),
|
||||
"ul=" + (n.language || n.browserLanguage || n.userLanguage),
|
||||
"sd=" + a.colorDepth + "-bit",
|
||||
"sr=" + a.width + "x" + a.height,
|
||||
"vp=" +
|
||||
g(m(4)) + '=' + g(m(6)),
|
||||
'ga=' + t.ga_tid,
|
||||
'dt=' + r(e.title),
|
||||
'de=' + r(e.characterSet || e.charset),
|
||||
'dr=' + r(e.referrer),
|
||||
'ul=' + (n.language || n.browserLanguage || n.userLanguage),
|
||||
'sd=' + a.colorDepth + '-bit',
|
||||
'sr=' + a.width + 'x' + a.height,
|
||||
'vp=' +
|
||||
o(e.documentElement.clientWidth, t.innerWidth || 0) +
|
||||
"x" +
|
||||
'x' +
|
||||
o(e.documentElement.clientHeight, t.innerHeight || 0),
|
||||
"plt=" + c(d.loadEventStart - d.navigationStart || 0),
|
||||
"dns=" + c(d.domainLookupEnd - d.domainLookupStart || 0),
|
||||
"pdt=" + c(d.responseEnd - d.responseStart || 0),
|
||||
"rrt=" + c(d.redirectEnd - d.redirectStart || 0),
|
||||
"tcp=" + c(d.connectEnd - d.connectStart || 0),
|
||||
"srt=" + c(d.responseStart - d.requestStart || 0),
|
||||
"dit=" + c(d.domInteractive - d.domLoading || 0),
|
||||
"clt=" + c(d.domContentLoadedEventStart - d.navigationStart || 0),
|
||||
"z=" + Date.now(),
|
||||
'plt=' + c(d.loadEventStart - d.navigationStart || 0),
|
||||
'dns=' + c(d.domainLookupEnd - d.domainLookupStart || 0),
|
||||
'pdt=' + c(d.responseEnd - d.responseStart || 0),
|
||||
'rrt=' + c(d.redirectEnd - d.redirectStart || 0),
|
||||
'tcp=' + c(d.connectEnd - d.connectStart || 0),
|
||||
'srt=' + c(d.responseStart - d.requestStart || 0),
|
||||
'dit=' + c(d.domInteractive - d.domLoading || 0),
|
||||
'clt=' + c(d.domContentLoadedEventStart - d.navigationStart || 0),
|
||||
'z=' + Date.now(),
|
||||
];
|
||||
(t.__ga_img = new Image()), (t.__ga_img.src = t.ga_api + "?" + i.join("&"));
|
||||
((t.__ga_img = new Image()), (t.__ga_img.src = t.ga_api + '?' + i.join('&')));
|
||||
}
|
||||
(t.cfga = s),
|
||||
"complete" === e.readyState ? s() : t.addEventListener("load", s);
|
||||
((t.cfga = s), 'complete' === e.readyState ? s() : t.addEventListener('load', s));
|
||||
})(window, document, navigator);
|
||||
|
@@ -8,12 +8,6 @@
|
||||
"baseUrl": "src",
|
||||
"noEmit": true
|
||||
},
|
||||
"include": [
|
||||
"src/**/*.ts",
|
||||
"src/**/*.tsx"
|
||||
],
|
||||
"exclude": [
|
||||
"node_modules",
|
||||
"dist"
|
||||
]
|
||||
"include": ["src/**/*.ts", "src/**/*.tsx"],
|
||||
"exclude": ["node_modules", "dist"]
|
||||
}
|
||||
|
@@ -1,7 +1,9 @@
|
||||
# open-gsio
|
||||
|
||||
[](https://github.com/geoffsee/open-gsio/actions/workflows/test.yml)
|
||||
[](https://opensource.org/licenses/MIT)
|
||||
</br>
|
||||
|
||||
<p align="center">
|
||||
<img src="https://github.com/user-attachments/assets/620d2517-e7be-4bb0-b2b7-3aa0cba37ef0" width="250" />
|
||||
</p>
|
||||
@@ -25,15 +27,15 @@
|
||||
- [License](#license)
|
||||
|
||||
## Stack
|
||||
* [TypeScript](https://www.typescriptlang.org/)
|
||||
* [Vike](https://vike.dev/)
|
||||
* [React](https://react.dev/)
|
||||
* [Cloudflare Workers](https://developers.cloudflare.com/workers/)
|
||||
* [itty‑router](https://github.com/kwhitley/itty-router)
|
||||
* [MobX‑State‑Tree](https://mobx-state-tree.js.org/)
|
||||
* [OpenAI SDK](https://github.com/openai/openai-node)
|
||||
* [Vitest](https://vitest.dev/)
|
||||
|
||||
- [TypeScript](https://www.typescriptlang.org/)
|
||||
- [Vike](https://vike.dev/)
|
||||
- [React](https://react.dev/)
|
||||
- [Cloudflare Workers](https://developers.cloudflare.com/workers/)
|
||||
- [itty‑router](https://github.com/kwhitley/itty-router)
|
||||
- [MobX‑State‑Tree](https://mobx-state-tree.js.org/)
|
||||
- [OpenAI SDK](https://github.com/openai/openai-node)
|
||||
- [Vitest](https://vitest.dev/)
|
||||
|
||||
## Installation
|
||||
|
||||
@@ -44,19 +46,22 @@
|
||||
> Note: it should be possible to use pnpm in place of bun.
|
||||
|
||||
## Deployment
|
||||
|
||||
1. Setup the KV_STORAGE bindings in `wrangler.jsonc`
|
||||
1. [Add another `GROQ_API_KEY` in secrets.json](https://console.groq.com/keys)
|
||||
1. Run `bun run deploy && bun run deploy:secrets && bun run deploy`
|
||||
|
||||
> Note: Subsequent deployments should omit `bun run deploy:secrets`
|
||||
|
||||
|
||||
## Local Inference
|
||||
|
||||
> Local inference is achieved by overriding the `OPENAI_API_KEY` and `OPENAI_API_ENDPOINT` environment variables. See below.
|
||||
|
||||
### mlx-omni-server
|
||||
|
||||
(default) (Apple Silicon Only) - Use Ollama for other platforms.
|
||||
~~~bash
|
||||
|
||||
```bash
|
||||
# (prereq) install mlx-omni-server
|
||||
brew tap seemueller-io/tap
|
||||
brew install seemueller-io/tap/mlx-omni-server
|
||||
@@ -64,10 +69,11 @@ brew install seemueller-io/tap/mlx-omni-server
|
||||
bun run openai:local mlx-omni-server # Start mlx-omni-server
|
||||
bun run openai:local:enable # Configure connection
|
||||
bun run server:dev # Restart server
|
||||
~~~
|
||||
```
|
||||
|
||||
#### Adding models for local inference (Apple Silicon)
|
||||
|
||||
~~~bash
|
||||
```bash
|
||||
# ensure mlx-omni-server is running
|
||||
|
||||
# See https://huggingface.co/mlx-community for available models
|
||||
@@ -79,22 +85,23 @@ curl http://localhost:10240/v1/chat/completions \
|
||||
\"model\": \"$MODEL_TO_ADD\",
|
||||
\"messages\": [{\"role\": \"user\", \"content\": \"Hello\"}]
|
||||
}"
|
||||
~~~
|
||||
```
|
||||
|
||||
### Ollama
|
||||
~~~bash
|
||||
|
||||
```bash
|
||||
bun run openai:local ollama # Start ollama server
|
||||
bun run openai:local:enable # Configure connection
|
||||
bun run server:dev # Restart server
|
||||
~~~
|
||||
```
|
||||
|
||||
#### Adding models for local inference (ollama)
|
||||
|
||||
~~~bash
|
||||
```bash
|
||||
# See https://ollama.com/library for available models
|
||||
MODEL_TO_ADD=gemma3
|
||||
docker exec -it ollama ollama run ${MODEL_TO_ADD}
|
||||
~~~
|
||||
|
||||
```
|
||||
|
||||
## Testing
|
||||
|
||||
@@ -102,20 +109,21 @@ Tests are located in `__tests__` directories next to the code they test. Testing
|
||||
|
||||
> `bun run test` will run all tests
|
||||
|
||||
|
||||
## Troubleshooting
|
||||
|
||||
1. `bun run clean`
|
||||
1. `bun i`
|
||||
1. `bun server:dev`
|
||||
1. `bun client:dev`
|
||||
1. Submit an issue
|
||||
|
||||
History
|
||||
---
|
||||
## History
|
||||
|
||||
A high-level overview for the development history of the parent repository, [geoff-seemueller-io](https://geoff.seemueller.io), is provided in [LEGACY.md](../../LEGACY.md).
|
||||
|
||||
## License
|
||||
~~~text
|
||||
|
||||
```text
|
||||
MIT License
|
||||
|
||||
Copyright (c) 2025 Geoff Seemueller
|
||||
@@ -137,5 +145,4 @@ AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
|
||||
LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM,
|
||||
OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
|
||||
SOFTWARE.
|
||||
~~~
|
||||
|
||||
```
|
||||
|
@@ -8,11 +8,6 @@
|
||||
"outDir": "dist",
|
||||
"rootDir": "."
|
||||
},
|
||||
"include": [
|
||||
"*.ts"
|
||||
],
|
||||
"exclude": [
|
||||
"node_modules",
|
||||
"*.test.ts"
|
||||
]
|
||||
"include": ["*.ts"],
|
||||
"exclude": ["node_modules", "*.test.ts"]
|
||||
}
|
||||
|
3
packages/env/env.d.ts
vendored
3
packages/env/env.d.ts
vendored
@@ -4,7 +4,7 @@ interface Env {
|
||||
EMAIL_SERVICE: any;
|
||||
|
||||
// Durable Objects
|
||||
SERVER_COORDINATOR: import("packages/server/durable-objects/ServerCoordinator.ts");
|
||||
SERVER_COORDINATOR: import('packages/server/durable-objects/ServerCoordinator.ts');
|
||||
|
||||
// Handles serving static assets
|
||||
ASSETS: Fetcher;
|
||||
@@ -12,7 +12,6 @@ interface Env {
|
||||
// KV Bindings
|
||||
KV_STORAGE: KVNamespace;
|
||||
|
||||
|
||||
// Text/Secrets
|
||||
METRICS_HOST: string;
|
||||
OPENAI_API_ENDPOINT: string;
|
||||
|
9
packages/env/tsconfig.json
vendored
9
packages/env/tsconfig.json
vendored
@@ -4,11 +4,6 @@
|
||||
"outDir": "dist",
|
||||
"rootDir": "."
|
||||
},
|
||||
"include": [
|
||||
"*.ts",
|
||||
"*.d.ts"
|
||||
],
|
||||
"exclude": [
|
||||
"node_modules"
|
||||
]
|
||||
"include": ["*.ts", "*.d.ts"],
|
||||
"exclude": ["node_modules"]
|
||||
}
|
@@ -6,11 +6,6 @@
|
||||
"allowJs": true,
|
||||
"noEmit": false
|
||||
},
|
||||
"include": [
|
||||
"*.js",
|
||||
"*.ts"
|
||||
],
|
||||
"exclude": [
|
||||
"node_modules"
|
||||
]
|
||||
"include": ["*.js", "*.ts"],
|
||||
"exclude": ["node_modules"]
|
||||
}
|
||||
|
@@ -10,12 +10,6 @@
|
||||
"allowJs": true,
|
||||
"jsx": "react-jsx"
|
||||
},
|
||||
"include": [
|
||||
"**/*.ts",
|
||||
"**/*.tsx"
|
||||
],
|
||||
"exclude": [
|
||||
"node_modules",
|
||||
"dist"
|
||||
]
|
||||
"include": ["**/*.ts", "**/*.tsx"],
|
||||
"exclude": ["node_modules", "dist"]
|
||||
}
|
||||
|
6
packages/server/types.d.ts
vendored
6
packages/server/types.d.ts
vendored
@@ -1,5 +1,5 @@
|
||||
declare global {
|
||||
type ExecutionContext = any
|
||||
type Env = import("@open-gsio/env")
|
||||
type ExecutionContext = any;
|
||||
type Env = import('@open-gsio/env');
|
||||
}
|
||||
export type ExecutionContext = any
|
||||
export type ExecutionContext = any;
|
||||
|
Reference in New Issue
Block a user