vscode-ai-toolkit icon indicating copy to clipboard operation
vscode-ai-toolkit copied to clipboard

Custom Models cannot be used as evaluator.

Open LEOCAO6 opened this issue 5 months ago • 4 comments

Add custom model with url/v1/chat/completions may cause API error when it is used in evaluator. Here is the error msg: 2025-07-07 22:46:00 +0800 6388 execution ERROR 5/5 flow run failed, indexes: [0,1,2,3,4], exception of index 0: Execution failure in 'router_eval_2.call': (WrappedOpenAIError) OpenAI API hits NotFoundError:

How to use custom model as evaluator without internet access?

LEOCAO6 avatar Jul 07 '25 14:07 LEOCAO6

Hi, @LEOCAO6 would you mind share why you want to use custom model for evaluations? Because usually local hosted small language models don't produce good results for evaluation.

a1exwang avatar Jul 15 '25 06:07 a1exwang

Hi Alex @a1exwang , the major reason for custom model evaluation is compliance and data sensitivity. In some FSI or Automobile use cases, data must be kept in local. Besides, local LLM like DeepSeek R1/V3 can have similar capabilities of API services from hyperscaler. Thank you!

LEOCAO6 avatar Jul 15 '25 15:07 LEOCAO6

@LEOCAO6 Thanks for the clarification of your use case. We have added this feature request to our backlog.

a1exwang avatar Jul 16 '25 02:07 a1exwang

I believe this feature is quite important, since it gives the freedom to also choose custom models which might be internal and adhere to certain standards. I would love to see this!

ChristophRaab avatar Nov 18 '25 15:11 ChristophRaab