You are browsing unreleased documentation.
This guide walks you through setting up the AI Proxy plugin with Amazon Bedrock.
Prerequisites
Configure the AI Proxy plugin
- Create a service in Kong Gateway that will represent the Bedrock API:
curl -X POST http://localhost:8001/services \ --data "name=bedrock-service" \ --data "url=https://bedrock-runtime.us-east-1.amazonaws.com"
Be sure to replace
us-east-1
with the region for your environment. - Create a route that maps to the service you defined:
curl -i -X POST http://localhost:8001/routes \ --data "paths[]=/bedrock" \ --data "service.id=$(curl -s http://localhost:8001/services/bedrock-service | jq -r '.id')"
- Use the Kong Admin API to configure the AI Proxy Plugin to route requests to Amazon Bedrock:
curl -X POST http://localhost:8001/services/bedrock-service/plugins \ --header 'Content-Type: application/json' \ --header 'accept: application/json' \ --data '{ "name": "ai-proxy", "instance_name": "ai-proxy-bedrock", "config": { "route_type": "llm/v1/chat", "auth": { "allow_override": false, "aws_access_key_id": "<YOUR_AWS_ACCESS_KEY_ID>", "aws_secret_access_key": "<YOUR_AWS_SECRET_ACCESS_KEY>" }, "model": { "provider": "bedrock", "name": "meta.llama3-70b-instruct-v1:0", "options": { "bedrock": { "aws_region": "us-east-1" } } } } }'
Be sure to replace the following values:
YOUR_AWS_ACCESS_KEY_ID
YOUR_AWS_SECRET_ACCESS_KEY
- Configure the correct provider model (
config.model.name
) and AWS region (config.model.options.bedrock.aws_region
) for your environment.
Test the configuration
Make an llm/v1/chat
type request to test your new endpoint:
curl -X POST http://localhost:8000/bedrock \
-H 'Content-Type: application/json' \
--data-raw '{ "messages": [ { "role": "system", "content": "You are a mathematician" }, { "role": "user", "content": "What is 1+1?"} ] }'