llama.cpp icon indicating copy to clipboard operation
llama.cpp copied to clipboard

Misc. bug: server not exit after `missing result_output tensor` error

Open ngxson opened this issue 1 week ago • 0 comments

Name and Version

While testing the rerank model on HF inference endpoint, we got this error: GGML_ASSERT(strcmp(res->name, "result_output") == 0 && "missing result_output tensor") failed

This is due to missing LLAMA_ARG_RERANKING (for reranking model) or LLAMA_ARG_EMBEDDINGS (for embeddings model).

The application is expected to edit after this error, but it still running which makes it a bit confused for end user.

Image

Expected behavior: the server should exit once it get that error.

Operating systems

No response

Which llama.cpp modules do you know to be affected?

No response

Command line

llama-server -m jina-rerank.gguf (do not add --rerank argument)

Problem description & steps to reproduce

Run a jina-rerank model without --rerank flag

First Bad Commit

No response

Relevant log output

(as seen on screenshot)

ngxson avatar Feb 11 '25 13:02 ngxson