Skip to content

Commit e6eb4cd

Browse files
dylankycfffonion
authored andcommitted
fix(ai-proxy): fix huggingface inference provider does not support use_cache and wait_for_model options [aigw-only] (#14287)
AG-506
1 parent 5f11809 commit e6eb4cd

File tree

2 files changed

+4
-47
lines changed

2 files changed

+4
-47
lines changed
Lines changed: 4 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,4 @@
1+
message: |
2+
**AI Proxy**: Fixed an issue where extra inputs were not permitted for huggingface inference provider
3+
type: "bugfix"
4+
scope: "Plugin"

kong/llm/drivers/huggingface.lua

Lines changed: 0 additions & 47 deletions
Original file line numberDiff line numberDiff line change
@@ -68,49 +68,6 @@ local function from_huggingface(response_string, model_info, route_type)
6868
return result_string, nil
6969
end
7070

71-
local function set_huggingface_options(model_info)
72-
local use_cache = false
73-
local wait_for_model = false
74-
75-
if model_info and model_info.options and model_info.options.huggingface then
76-
use_cache = model_info.options.huggingface.use_cache or false
77-
wait_for_model = model_info.options.huggingface.wait_for_model or false
78-
end
79-
80-
return {
81-
use_cache = use_cache,
82-
wait_for_model = wait_for_model,
83-
}
84-
end
85-
86-
local function set_default_parameters(request_table)
87-
local parameters = request_table.parameters or {}
88-
if parameters.top_k == nil then
89-
parameters.top_k = request_table.top_k
90-
end
91-
if parameters.top_p == nil then
92-
parameters.top_p = request_table.top_p
93-
end
94-
if parameters.temperature == nil then
95-
parameters.temperature = request_table.temperature
96-
end
97-
if parameters.max_tokens == nil then
98-
if request_table.messages then
99-
-- conversational model use the max_length param
100-
-- https://huggingface.co/docs/api-inference/en/detailed_parameters?code=curl#conversational-task
101-
parameters.max_length = request_table.max_tokens
102-
else
103-
parameters.max_new_tokens = request_table.max_tokens
104-
end
105-
end
106-
request_table.top_k = nil
107-
request_table.top_p = nil
108-
request_table.temperature = nil
109-
request_table.max_tokens = nil
110-
111-
return parameters
112-
end
113-
11471
local function safe_access(tbl, ...)
11572
local value = tbl
11673
for _, key in ipairs({ ... }) do
@@ -195,19 +152,15 @@ local function to_huggingface(task, request_table, model_info)
195152
if task == "llm/v1/completions" then
196153
request_table.inputs = request_table.prompt
197154
request_table.prompt = nil
198-
request_table.parameters = set_default_parameters(request_table)
199155
request_table.model = model_info.name or request_table.model
200156

201157
elseif task == "llm/v1/chat" then
202158
-- For router.huggingface.co, we need to include the model in the request body
203159
request_table.model = model_info.name or request_table.model
204-
request_table.parameters = set_default_parameters(request_table)
205160
request_table.inputs = request_table.prompt
206161

207162
end
208163

209-
request_table.options = set_huggingface_options(model_info)
210-
211164
return request_table, "application/json", nil
212165
end
213166

0 commit comments

Comments
 (0)