k8sgpt
k8sgpt copied to clipboard
[Feature]: backend support for Hugging Face
Checklist
- [X] I've searched for similar issues and couldn't find anything matching
- [X] I've discussed this feature request in the K8sGPT Slack and got positive feedback
Is this feature request related to a problem?
None
Problem Description
No response
Solution Description
backend support for HF models. users can use inference API of HF conversational model.
for HF interface, there are some packages like https://pkg.go.dev/github.com/hupe1980/go-huggingface
Benefits
using hf backend, k8sgpt users can use free llm api through hf. also, sLLMs are good enought on k8s analyzing.
Potential Drawbacks
- Inference API on HF is not for production. it is kind of serverless. so sometimes api responses like
huggingfaces error: Model mistralai/Mistral-7B-v0.1 is currently loading
. So hf backend should be used to locally or when testing llm model. - not all models in hf are available. some models` inference API is inactivated, and even some are not work even in HF page. (below screenshot)
Additional Information
No response
for go huggingface client packages (https://github.com/hupe1980/go-huggingface, https://github.com/Kardbord/hfapigo), they both are http request wrapper for HF inference API. which means it’s go version of this python code. since what we need is just conversation API, and their community is not strong, I'm not sure about using these packages.
import requests
API_URL = "https://api-inference.huggingface.co/models/microsoft/DialoGPT-medium"
headers = {"Authorization": "Bearer ###"}
def query(payload):
response = requests.post(API_URL, headers=headers, json=payload)
return response.json()
output = query({
"inputs": {
"past_user_inputs": ["Which movie is the best ?"],
"generated_responses": ["It is Die Hard for sure."],
"text": "Can you explain why ?"
},
})