@@ -68,49 +68,6 @@ local function from_huggingface(response_string, model_info, route_type)
6868 return result_string , nil
6969end
7070
71- local function set_huggingface_options (model_info )
72- local use_cache = false
73- local wait_for_model = false
74-
75- if model_info and model_info .options and model_info .options .huggingface then
76- use_cache = model_info .options .huggingface .use_cache or false
77- wait_for_model = model_info .options .huggingface .wait_for_model or false
78- end
79-
80- return {
81- use_cache = use_cache ,
82- wait_for_model = wait_for_model ,
83- }
84- end
85-
86- local function set_default_parameters (request_table )
87- local parameters = request_table .parameters or {}
88- if parameters .top_k == nil then
89- parameters .top_k = request_table .top_k
90- end
91- if parameters .top_p == nil then
92- parameters .top_p = request_table .top_p
93- end
94- if parameters .temperature == nil then
95- parameters .temperature = request_table .temperature
96- end
97- if parameters .max_tokens == nil then
98- if request_table .messages then
99- -- conversational model use the max_length param
100- -- https://huggingface.co/docs/api-inference/en/detailed_parameters?code=curl#conversational-task
101- parameters .max_length = request_table .max_tokens
102- else
103- parameters .max_new_tokens = request_table .max_tokens
104- end
105- end
106- request_table .top_k = nil
107- request_table .top_p = nil
108- request_table .temperature = nil
109- request_table .max_tokens = nil
110-
111- return parameters
112- end
113-
11471local function safe_access (tbl , ...)
11572 local value = tbl
11673 for _ , key in ipairs ({ ... }) do
@@ -195,19 +152,15 @@ local function to_huggingface(task, request_table, model_info)
195152 if task == " llm/v1/completions" then
196153 request_table .inputs = request_table .prompt
197154 request_table .prompt = nil
198- request_table .parameters = set_default_parameters (request_table )
199155 request_table .model = model_info .name or request_table .model
200156
201157 elseif task == " llm/v1/chat" then
202158 -- For router.huggingface.co, we need to include the model in the request body
203159 request_table .model = model_info .name or request_table .model
204- request_table .parameters = set_default_parameters (request_table )
205160 request_table .inputs = request_table .prompt
206161
207162 end
208163
209- request_table .options = set_huggingface_options (model_info )
210-
211164 return request_table , " application/json" , nil
212165end
213166
0 commit comments