An OpenAI api wrapper for the Runpod vllm docker image
OpenAI api reference
- Supports Vicuna and Llama 2 prompt format
- Only supports streaming for now
- Works with SillyTavern as OpenAI proxy
Edit src/constants.ts (Recommended to add a SECRET_TOKEN for safety) Running locally:
pnpm install
pnpm run dev
Uploading to Cloudflare Worker
pnpm run deploy