Access your local AI models using standard OpenAI endpoints.
The service is available at http://localhost:8080. You can use standard HTTP requests to interact with it.
curl http://localhost:8080/v1/chat/completions \
-H "Content-Type: application/json" \
-d '{
"model": "gpt-3.5-turbo",
"messages": [{"role": "user", "content": "Hello!"}]
}'
/v1/chat/completions
/v1/images/generations
/v1/audio/speech