Skip to content

kirel/raycast-openai-server

Repository files navigation

README

Development

npm run dev

curl -N -X POST http://localhost:1235/v1/chat/completions \
  -H "Content-Type: application/json" \
  -d '{"model": "google-gemini-2.0-flash", "stream": true, "messages": [{"role": "user", "content": "Who is the president?"}]}'

curl http://localhost:1235/v1/models | jq .

curl -X POST http://localhost:1235/kill

Aider

Bootstrap with raycast AI chat window

OPENAI_API_BASE=http://localhost:1234/v1 aider --copy-paste --model openai/mistral-small-24b-instruct-2501

Eating your own dogfood - use OpenAI server

OPENAI_API_BASE=http://localhost:1235/v1 aider --model openai/together-deepseek-ai/DeepSeek-R1
OPENAI_API_BASE=http://localhost:1235/v1 aider --model openai/openai_o1-o3-mini
OPENAI_API_BASE=http://localhost:1235/v1 aider --watch-files --model openai/anthropic-claude-sonnet

Add stuff to aider

/read-only README.md
/add src/*.tsx src/*.ts package.json
https://developers.raycast.com/api-reference/menu-bar-commands

Goal

Create a Raycast Extension that provides an OpenAI compatible API server

About

Extension that thats local openai server which passes requests to Raycast AI.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published