blob: ccc84adc71413e887996c860876a233001198c2d [file] [log] [blame]
#
# Licensed to the Apache Software Foundation (ASF) under one or more
# contributor license agreements. See the NOTICE file distributed with
# this work for additional information regarding copyright ownership.
# The ASF licenses this file to You under the Apache License, Version 2.0
# (the "License"); you may not use this file except in compliance with
# the License. You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
#
use t::APISIX 'no_plan';
log_level("info");
repeat_each(1);
no_long_string();
no_root_location();
my $resp_file = 't/assets/ai-proxy-response.json';
open(my $fh, '<', $resp_file) or die "Could not open file '$resp_file' $!";
my $resp = do { local $/; <$fh> };
close($fh);
print "Hello, World!\n";
print $resp;
add_block_preprocessor(sub {
my ($block) = @_;
if (!defined $block->request) {
$block->set_value("request", "GET /t");
}
my $user_yaml_config = <<_EOC_;
plugins:
- ai-proxy-multi
- prometheus
_EOC_
$block->set_value("extra_yaml_config", $user_yaml_config);
my $http_config = $block->http_config // <<_EOC_;
server {
server_name openai_rate_limit;
default_type 'application/json';
listen 6726;
location / {
content_by_lua_block {
ngx.status = 429
ngx.say([[{ "error": {"message":"rate limit exceeded"}}]])
return
}
}
}
server {
server_name openai_internal_error;
default_type 'application/json';
listen 6727;
location / {
content_by_lua_block {
ngx.status = 500
ngx.say([[{ "error": {"message":"internal server error"}}]])
return
}
}
}
server {
server_name openai_internal_error;
default_type 'application/json';
listen 6728;
location / {
content_by_lua_block {
ngx.status = 503
ngx.say([[{ "error": {"message":"service unavailable"}}]])
return
}
}
}
server {
server_name openai;
listen 6724;
default_type 'application/json';
location /v1/chat/completions {
content_by_lua_block {
local json = require("cjson.safe")
if ngx.req.get_method() ~= "POST" then
ngx.status = 400
ngx.say("Unsupported request method: ", ngx.req.get_method())
end
ngx.req.read_body()
local body, err = ngx.req.get_body_data()
body, err = json.decode(body)
local header_auth = ngx.req.get_headers()["authorization"]
local query_auth = ngx.req.get_uri_args()["apikey"]
if header_auth ~= "Bearer token" and query_auth ~= "apikey" then
ngx.status = 401
ngx.say("Unauthorized")
return
end
if header_auth == "Bearer token" or query_auth == "apikey" then
ngx.req.read_body()
local body, err = ngx.req.get_body_data()
body, err = json.decode(body)
if not body.messages or #body.messages < 1 then
ngx.status = 400
ngx.say([[{ "error": "bad request"}]])
return
end
ngx.status = 200
ngx.print("openai")
return
end
ngx.status = 503
ngx.say("reached the end of the test suite")
}
}
location /chat/completions {
content_by_lua_block {
local json = require("cjson.safe")
if ngx.req.get_method() ~= "POST" then
ngx.status = 400
ngx.say("Unsupported request method: ", ngx.req.get_method())
end
ngx.req.read_body()
local body, err = ngx.req.get_body_data()
body, err = json.decode(body)
local header_auth = ngx.req.get_headers()["authorization"]
local query_auth = ngx.req.get_uri_args()["apikey"]
if header_auth ~= "Bearer token" and query_auth ~= "apikey" then
ngx.status = 401
ngx.say("Unauthorized")
return
end
if header_auth == "Bearer token" or query_auth == "apikey" then
ngx.req.read_body()
local body, err = ngx.req.get_body_data()
body, err = json.decode(body)
if not body.messages or #body.messages < 1 then
ngx.status = 400
ngx.say([[{ "error": "bad request"}]])
return
end
ngx.status = 200
ngx.print("deepseek")
return
end
ngx.status = 503
ngx.say("reached the end of the test suite")
}
}
}
_EOC_
$block->set_value("http_config", $http_config);
});
run_tests();
__DATA__
=== TEST 1: set route with roundrobin balancer, weight 4 and 1
--- config
location /t {
content_by_lua_block {
local t = require("lib.test_admin").test
local code, body = t('/apisix/admin/routes/1',
ngx.HTTP_PUT,
[[{
"uri": "/anything",
"plugins": {
"ai-proxy-multi": {
"instances": [
{
"name": "openai",
"provider": "openai",
"weight": 4,
"auth": {
"header": {
"Authorization": "Bearer token"
}
},
"options": {
"model": "gpt-4",
"max_tokens": 512,
"temperature": 1.0
},
"override": {
"endpoint": "http://localhost:6724"
}
},
{
"name": "deepseek",
"provider": "deepseek",
"weight": 1,
"auth": {
"header": {
"Authorization": "Bearer token"
}
},
"options": {
"model": "deepseek-chat",
"max_tokens": 512,
"temperature": 1.0
},
"override": {
"endpoint": "http://localhost:6724/chat/completions"
}
}
],
"ssl_verify": false
}
}
}]]
)
if code >= 300 then
ngx.status = code
end
ngx.say(body)
}
}
--- response_body
passed
=== TEST 2: test
--- config
location /t {
content_by_lua_block {
local http = require "resty.http"
local uri = "http://127.0.0.1:" .. ngx.var.server_port
.. "/anything"
local restab = {}
local body = [[{ "messages": [ { "role": "system", "content": "You are a mathematician" }, { "role": "user", "content": "What is 1+1?"} ] }]]
for i = 1, 10 do
local httpc = http.new()
local res, err = httpc:request_uri(uri, {method = "POST", body = body})
if not res then
ngx.say(err)
return
end
table.insert(restab, res.body)
end
table.sort(restab)
ngx.log(ngx.WARN, "test picked instances: ", table.concat(restab, "."))
}
}
--- request
GET /t
--- error_log
deepseek.deepseek.openai.openai.openai.openai.openai.openai.openai.openai
=== TEST 3: set route with chash balancer, weight 4 and 1
--- config
location /t {
content_by_lua_block {
local t = require("lib.test_admin").test
local code, body = t('/apisix/admin/routes/1',
ngx.HTTP_PUT,
[[{
"uri": "/anything",
"plugins": {
"ai-proxy-multi": {
"balancer": {
"algorithm": "chash",
"hash_on": "vars",
"key": "query_string"
},
"instances": [
{
"name": "openai",
"provider": "openai",
"weight": 4,
"auth": {
"header": {
"Authorization": "Bearer token"
}
},
"options": {
"model": "gpt-4",
"max_tokens": 512,
"temperature": 1.0
},
"override": {
"endpoint": "http://localhost:6724"
}
},
{
"name": "deepseek",
"provider": "deepseek",
"weight": 1,
"auth": {"header": {"Authorization": "Bearer token"}},
"options": {"model": "deepseek-chat","max_tokens": 512,"temperature": 1.0},
"override": {"endpoint": "http://localhost:6724/chat/completions"}
}
],
"ssl_verify": false
}
}
}]]
)
if code >= 300 then
ngx.status = code
end
ngx.say(body)
}
}
--- response_body
passed
=== TEST 4: test
--- config
location /t {
content_by_lua_block {
local http = require "resty.http"
local uri = "http://127.0.0.1:" .. ngx.var.server_port
.. "/anything"
local restab = {}
local body = [[{ "messages": [ { "role": "system", "content": "You are a mathematician" }, { "role": "user", "content": "What is 1+1?"} ] }]]
for i = 1, 10 do
local httpc = http.new()
local query = {
index = i
}
local res, err = httpc:request_uri(uri, {method = "POST", body = body, query = query})
if not res then
ngx.say(err)
return
end
table.insert(restab, res.body)
end
local count = {}
for _, value in ipairs(restab) do
count[value] = (count[value] or 0) + 1
end
for p, num in pairs(count) do
ngx.log(ngx.WARN, "distribution: ", p, ": ", num)
end
}
}
--- request
GET /t
--- timeout: 10
--- error_log
distribution: deepseek: 2
distribution: openai: 8
=== TEST 5: set route with fallback_strategy with 500 response code openai.
--- config
location /t {
content_by_lua_block {
local t = require("lib.test_admin").test
local code, body = t('/apisix/admin/routes/1',
ngx.HTTP_PUT,
[[{
"uri": "/anything",
"plugins": {
"ai-proxy-multi": {
"fallback_strategy": ["http_5xx"],
"balancer": {
"algorithm": "chash",
"hash_on": "vars",
"key": "query_string"
},
"instances": [
{
"name": "openai",
"provider": "openai",
"weight": 4,
"auth": {
"header": {
"Authorization": "Bearer token"
}
},
"options": {
"model": "gpt-4",
"max_tokens": 512,
"temperature": 1.0
},
"override": {
"endpoint": "http://localhost:6727"
}
},
{
"name": "deepseek",
"provider": "deepseek",
"weight": 1,
"auth": {"header": {"Authorization": "Bearer token"}},
"options": {"model": "deepseek-chat","max_tokens": 512,"temperature": 1.0},"override": {"endpoint": "http://localhost:6724/chat/completions"}}
],
"ssl_verify": false
}
}
}]]
)
if code >= 300 then
ngx.status = code
end
ngx.say(body)
}
}
--- response_body
passed
=== TEST 6: test all requests success with fallback deepseek
--- config
location /t {
content_by_lua_block {
local http = require "resty.http"
local uri = "http://127.0.0.1:" .. ngx.var.server_port
.. "/anything"
local restab = {}
local body = [[{ "messages": [ { "role": "system", "content": "You are a mathematician" }, { "role": "user", "content": "What is 1+1?"} ] }]]
for i = 1, 10 do
local httpc = http.new()
local query = {
index = i
}
local res, err = httpc:request_uri(uri, {method = "POST", body = body, query = query})
if not res then
ngx.say(err)
return
end
table.insert(restab, res.body)
end
local count = {}
for _, value in ipairs(restab) do
count[value] = (count[value] or 0) + 1
end
for p, num in pairs(count) do
ngx.log(ngx.WARN, "distribution: ", p, ": ", num)
end
}
}
--- request
GET /t
--- timeout: 10
--- error_log
distribution: deepseek: 10
=== TEST 7: set route with fallback_strategy with too many requests openai.
--- config
location /t {
content_by_lua_block {
local t = require("lib.test_admin").test
local code, body = t('/apisix/admin/routes/1',
ngx.HTTP_PUT,
[[{
"uri": "/anything",
"plugins": {
"ai-proxy-multi": {
"fallback_strategy": ["http_429"],
"balancer": {
"algorithm": "chash",
"hash_on": "vars",
"key": "query_string"
},
"instances": [
{
"name": "openai",
"provider": "openai",
"weight": 4,
"auth": {
"header": {
"Authorization": "Bearer token"
}
},
"options": {
"model": "gpt-4",
"max_tokens": 512,
"temperature": 1.0
},
"override": {
"endpoint": "http://localhost:6726"
}
},
{"name":"deepseek","provider":"deepseek","weight":1,"auth":{"header":{"Authorization":"Bearer token"}},"options":{"model":"deepseek-chat","max_tokens":512,"temperature":1.0},"override":{"endpoint":"http://localhost:6724/chat/completions"}}
],
"ssl_verify": false
}
}
}]]
)
if code >= 300 then
ngx.status = code
end
ngx.say(body)
}
}
--- response_body
passed
=== TEST 8: test all requests success with fallback deepseek
--- config
location /t {
content_by_lua_block {
local http = require "resty.http"
local uri = "http://127.0.0.1:" .. ngx.var.server_port
.. "/anything"
local restab = {}
local body = [[{ "messages": [ { "role": "system", "content": "You are a mathematician" }, { "role": "user", "content": "What is 1+1?"} ] }]]
for i = 1, 10 do
local httpc = http.new()
local query = {
index = i
}
local res, err = httpc:request_uri(uri, {method = "POST", body = body, query = query})
if not res then
ngx.say(err)
return
end
table.insert(restab, res.body)
end
local count = {}
for _, value in ipairs(restab) do
count[value] = (count[value] or 0) + 1
end
for p, num in pairs(count) do
ngx.log(ngx.WARN, "distribution: ", p, ": ", num)
end
}
}
--- request
GET /t
--- timeout: 10
--- error_log
distribution: deepseek: 10
=== TEST 9: set route with fallback_strategy with unreachable openai.
--- config
location /t {
content_by_lua_block {
local t = require("lib.test_admin").test
local code, body = t('/apisix/admin/routes/1',
ngx.HTTP_PUT,
[[{
"uri": "/anything",
"plugins": {
"ai-proxy-multi": {
"fallback_strategy": ["http_5xx"],
"balancer": {
"algorithm": "chash",
"hash_on": "vars",
"key": "query_string"
},
"instances": [
{
"name": "openai",
"provider": "openai",
"weight": 4,
"auth": {
"header": {
"Authorization": "Bearer token"
}
},
"options": {
"model": "gpt-4",
"max_tokens": 512,
"temperature": 1.0
},
"override": {
"endpoint": "http://localhost:6725"
}
},
{"name":"deepseek","provider":"deepseek","weight":1,"auth":{"header":{"Authorization":"Bearer token"}},"options":{"model":"deepseek-chat","max_tokens":512,"temperature":1.0},"override":{"endpoint":"http://localhost:6724/chat/completions"}}
],
"ssl_verify": false
}
}
}]]
)
if code >= 300 then
ngx.status = code
end
ngx.say(body)
}
}
--- response_body
passed
=== TEST 10: test all requests success with fallback deepseek
--- config
location /t {
content_by_lua_block {
local http = require "resty.http"
local uri = "http://127.0.0.1:" .. ngx.var.server_port
.. "/anything"
local restab = {}
local body = [[{ "messages": [ { "role": "system", "content": "You are a mathematician" }, { "role": "user", "content": "What is 1+1?"} ] }]]
for i = 1, 10 do
local httpc = http.new()
local query = {
index = i
}
local res, err = httpc:request_uri(uri, {method = "POST", body = body, query = query})
if not res then
ngx.say(err)
return
end
table.insert(restab, res.body)
end
local count = {}
for _, value in ipairs(restab) do
count[value] = (count[value] or 0) + 1
end
for p, num in pairs(count) do
ngx.log(ngx.WARN, "distribution: ", p, ": ", num)
end
}
}
--- request
GET /t
--- timeout: 10
--- error_log
distribution: deepseek: 10
=== TEST 11: set route with fallback_strategy with service unavailable openai.
--- config
location /t {
content_by_lua_block {
local t = require("lib.test_admin").test
local code, body = t('/apisix/admin/routes/1',
ngx.HTTP_PUT,
[[{
"uri": "/anything",
"plugins": {
"ai-proxy-multi": {
"fallback_strategy": ["http_5xx"],
"balancer": {
"algorithm": "chash",
"hash_on": "vars",
"key": "query_string"
},
"instances": [
{
"name": "openai",
"provider": "openai",
"weight": 4,
"auth": {
"header": {
"Authorization": "Bearer token"
}
},
"options": {
"model": "gpt-4",
"max_tokens": 512,
"temperature": 1.0
},
"override": {
"endpoint": "http://localhost:6728"
}
},
{"name":"deepseek","provider":"deepseek","weight":1,"auth":{"header":{"Authorization":"Bearer token"}},"options":{"model":"deepseek-chat","max_tokens":512,"temperature":1.0},"override":{"endpoint":"http://localhost:6724/chat/completions"}}
],
"ssl_verify": false
}
}
}]]
)
if code >= 300 then
ngx.status = code
end
ngx.say(body)
}
}
--- response_body
passed
=== TEST 12: test all requests success with fallback deepseek
--- config
location /t {
content_by_lua_block {
local http = require "resty.http"
local uri = "http://127.0.0.1:" .. ngx.var.server_port
.. "/anything"
local restab = {}
local body = [[{ "messages": [ { "role": "system", "content": "You are a mathematician" }, { "role": "user", "content": "What is 1+1?"} ] }]]
for i = 1, 10 do
local httpc = http.new()
local query = {
index = i
}
local res, err = httpc:request_uri(uri, {method = "POST", body = body, query = query})
if not res then
ngx.say(err)
return
end
table.insert(restab, res.body)
end
local count = {}
for _, value in ipairs(restab) do
count[value] = (count[value] or 0) + 1
end
for p, num in pairs(count) do
ngx.log(ngx.WARN, "distribution: ", p, ": ", num)
end
}
}
--- request
GET /t
--- timeout: 10
--- error_log
distribution: deepseek: 10
=== TEST 13: set route with fallback_strategy with service unavailable openai having high priority.
--- config
location /t {
content_by_lua_block {
local t = require("lib.test_admin").test
local code, body = t('/apisix/admin/routes/1',
ngx.HTTP_PUT,
[[{
"uri": "/anything",
"plugins": {
"ai-proxy-multi": {
"balancer": {
"algorithm": "roundrobin",
"hash_on": "vars"
},
"fallback_strategy": [
"http_429",
"http_5xx"
],
"instances": [
{"auth":{"header":{"Authorization":"Bearer token"}},"name":"mock-429","override":{"endpoint":"http://localhost:6726"},"priority":10,"provider":"openai-compatible","weight":10},{"auth":{"header":{"Authorization":"Bearer token"}},"name":"mock-500","override":{"endpoint":"http://localhost:6727"},"priority":0,"provider":"openai-compatible","weight":10},{"auth":{"header":{"Authorization":"Bearer token"}},"name":"mock-200","override":{"endpoint":"http://localhost:6724/chat/completions"},"priority":0,"provider":"openai-compatible","weight":1}
],
"ssl_verify": false
}
}
}]]
)
if code >= 300 then
ngx.status = code
end
ngx.say(body)
}
}
--- response_body
passed
=== TEST 14: test all requests success with fallback deepseek
--- config
location /t {
content_by_lua_block {
local http = require "resty.http"
local uri = "http://127.0.0.1:" .. ngx.var.server_port
.. "/anything"
local restab = {}
local body = [[{ "messages": [ { "role": "system", "content": "You are a mathematician" }, { "role": "user", "content": "What is 1+1?"} ] }]]
for i = 1, 10 do
local httpc = http.new()
local query = {
index = i
}
local res, err = httpc:request_uri(uri, {method = "POST", body = body, query = query})
if not res then
ngx.say(err)
return
end
table.insert(restab, res.body)
end
local count = {}
for _, value in ipairs(restab) do
count[value] = (count[value] or 0) + 1
end
for p, num in pairs(count) do
ngx.log(ngx.WARN, "distribution: ", p, ": ", num)
end
}
}
--- request
GET /t
--- timeout: 10
--- error_log
distribution: deepseek: 10
=== TEST 15: set route with fallback_strategy with only service unavailable and 429.
--- config
location /t {
content_by_lua_block {
local t = require("lib.test_admin").test
local code, body = t('/apisix/admin/routes/1',
ngx.HTTP_PUT,
[[{
"uri": "/anything",
"plugins": {
"ai-proxy-multi": {
"balancer": {
"algorithm": "roundrobin",
"hash_on": "vars"
},
"fallback_strategy": [
"http_429",
"http_5xx"
],
"instances": [
{
"auth": {
"header": {
"Authorization": "Bearer token"
}
},
"name": "mock-429",
"override": {
"endpoint": "http://localhost:6726"
},
"priority": 10,
"provider": "openai-compatible",
"weight": 10
},
{
"auth": {
"header": {
"Authorization": "Bearer token"
}
},
"name": "mock-500",
"override": {
"endpoint": "http://localhost:6727"
},
"priority": 0,
"provider": "openai-compatible",
"weight": 10
}
],
"ssl_verify": false
}
}
}]]
)
if code >= 300 then
ngx.status = code
end
ngx.say(body)
}
}
--- response_body
passed
=== TEST 16: test all requests success with fallback deepseek should return 502
--- config
location /t {
content_by_lua_block {
local http = require "resty.http"
local uri = "http://127.0.0.1:" .. ngx.var.server_port
.. "/anything"
local restab = {}
local body = [[{ "messages": [ { "role": "system", "content": "You are a mathematician" }, { "role": "user", "content": "What is 1+1?"} ] }]]
for i = 1, 10 do
local httpc = http.new()
local query = {
index = i
}
local res, err = httpc:request_uri(uri, {method = "POST", body = body, query = query})
if not res then
ngx.say(err)
return
end
table.insert(restab, res.status)
end
local count = {}
for _, value in ipairs(restab) do
count[value] = (count[value] or 0) + 1
end
for p, num in pairs(count) do
ngx.log(ngx.WARN, "distribution: ", p, ": ", num)
end
}
}
--- request
GET /t
--- timeout: 10
--- error_log
distribution: 502: 10
=== TEST 17: set route with only one instance and configure it with http_5xx fallback_strategy
--- config
location /t {
content_by_lua_block {
local t = require("lib.test_admin").test
local json = require("toolkit.json")
local request_body = {
uri = "/anything",
plugins = {
["ai-proxy-multi"] = {
fallback_strategy = {"http_5xx"},
instances = {
{
name = "deepseek",
provider = "deepseek",
auth = {
header = {Authorization = "Bearer token" }
},
weight = 1,
override = { endpoint = "http://localhost:6727" }
}
},
ssl_verify = false
}
}
}
local code, body = t('/apisix/admin/routes/1',
ngx.HTTP_PUT,
json.encode(request_body)
)
if code >= 300 then
ngx.status = code
end
ngx.say(body)
}
}
--- response_body
passed
=== TEST 18: send request
--- request
POST /anything
{ "messages": [ { "role": "system", "content": "You are a mathematician" }, { "role": "user", "content": "What is 1+1?"} ] }
--- more_headers
Host: openai_internal_error
--- error_code: 500
--- no_error_log
[error]