Skip to content

Local Development with Docker/Podman

Run stdapi.ai locally for development, testing, and evaluation using the free community container image (AGPL-3.0). Full API compatibility — the same endpoints and features as the production deployment.

New to AWS Bedrock?

To run stdapi.ai locally you need:

  1. An AWS accountcreate one free
  2. AWS credentials configured locally — via aws configure or aws sso login (AWS CLI setup guide)

Run It

With AWS credentials (after aws sso login):

docker run --rm -p 8000:8000 \
  -v ~/.aws:/home/nonroot/.aws:ro \
  -e AWS_BEDROCK_REGIONS=us-east-1,us-west-2 \
  -e ENABLE_DOCS=true \
  ghcr.io/stdapi-ai/stdapi.ai-community:latest

With environment variables instead:

docker run --rm -p 8000:8000 \
  -e AWS_ACCESS_KEY_ID=your-access-key-id \
  -e AWS_SECRET_ACCESS_KEY=your-secret-access-key \
  -e AWS_SESSION_TOKEN=your-session-token \
  -e AWS_BEDROCK_REGIONS=us-east-1,us-west-2 \
  -e ENABLE_DOCS=true \
  ghcr.io/stdapi-ai/stdapi.ai-community:latest
Podman on Fedora/RHEL with SELinux

Add the :z SELinux label and --userns=keep-id:

podman run --rm -p 8000:8000 \
  --userns=keep-id \
  -v ~/.aws:/home/nonroot/.aws:ro,z \
  -e AWS_BEDROCK_REGIONS=us-east-1,us-west-2 \
  -e ENABLE_DOCS=true \
  ghcr.io/stdapi-ai/stdapi.ai-community:latest

The :z flag relabels files for container access. Use :Z if multiple containers share the volume. --userns=keep-id maps your host user ID to the container user.

%%{init: {'flowchart': {'htmlLabels': true}} }%%
flowchart LR
  openai["<img src='../styles/logo_openai.svg' style='height:64px;width:auto;vertical-align:middle;' /> OpenAI SDK"] --> local["<img src='../styles/logo.svg' style='height:64px;width:auto;vertical-align:middle;' /> stdapi.ai (community)<br/>Docker/Podman"]
  anthropic["<img src='../styles/logo_anthropic.svg' style='height:64px;width:auto;vertical-align:middle;' /> Anthropic SDK"] --> local
  local --> bedrock["<img src='../styles/logo_amazon_bedrock.svg' style='height:64px;width:auto;vertical-align:middle;' /> AWS Bedrock"]
  local --> polly["<img src='../styles/logo_amazon_polly.svg' style='height:64px;width:auto;vertical-align:middle;' /> AWS Polly"]
  local --> transcribe["<img src='../styles/logo_amazon_transcribe.svg' style='height:64px;width:auto;vertical-align:middle;' /> AWS Transcribe"]
  local --> s3["<img src='../styles/logo_amazon_s3.svg' style='height:64px;width:auto;vertical-align:middle;' /> AWS S3"]

Test It

# Check health
curl http://localhost:8000/health

# List all available models
curl http://localhost:8000/search_models

# Search models by capability (e.g. streaming-capable chat models only)
curl "http://localhost:8000/search_models?route=/v1/chat/completions&streaming=true"

# Chat completion
curl http://localhost:8000/v1/chat/completions \
  -H "Content-Type: application/json" \
  -d '{
    "model": "amazon.nova-micro-v1:0",
    "messages": [{"role": "user", "content": "Hello!"}]
  }'

Interactive API docs: Open http://localhost:8000/docs for Swagger UI with all available endpoints.

Try other models

amazon.nova-micro-v1:0 is a fast, low-cost model — great for confirming the pipeline works. Once you see a response, switch the model field to anthropic.claude-opus-4-7, anthropic.claude-sonnet-4-6, or any other Bedrock model available in your configured regions.

Default model discovery: GET /search_models (no parameters) returns every model the gateway has discovered, with full details — provider, input/output modalities, supported routes, regions, streaming and legacy status. Add query parameters to filter by capability (e.g. ?input_modalities=IMAGE&route=/v1/chat/completions for vision-capable chat models). This is also the recommended endpoint for AI agents to pick the right model ID before invoking another endpoint — see the Search Models API reference. GET /v1/models is also available for strict OpenAI SDK compatibility.


What to Try Next

stdapi.ai works with any OpenAI or Anthropic-compatible tool. Here are popular integrations to try locally:

  • Open WebUI — Private ChatGPT-like interface with RAG, multi-modal support, and document upload
  • n8n Workflows — AI-powered automation with 400+ integrations
  • AI Coding Assistants — Claude Code, Continue.dev, Cline, Cursor, Aider with AWS Bedrock models
  • API Overview — All endpoints, parameters, and usage examples
  • Troubleshooting — Podman/SELinux, auth, model-not-found, and other common errors

Technical Notes

Building from source: See the Dockerfile if you prefer to build the image yourself.

Container runtime: Both community and Marketplace images use Granian, a high-performance Python ASGI server. Granian environment variables (e.g., GRANIAN_PORT, GRANIAN_WORKERS) are supported.

Configuration: See Configuration Reference for all environment variables.

Ready for Production?

When you're ready to deploy to AWS with HTTPS, auto-scaling, and enterprise features, the production deployment guide gets you running in 5 minutes with Terraform. The AWS Marketplace subscription includes a 14-day free trial.