GPULlama3.java icon indicating copy to clipboard operation
GPULlama3.java copied to clipboard

Build GUI Chatbox for GPULlama3.java Inference with embedded GPU resource monitoring

Open mikepapadim opened this issue 6 months ago • 2 comments

Description:

We want to create a user-friendly chat-style GUI for running LLM inference via our Java engine (powered by TornadoVM), inspired by the layout shown in the attached screenshot.

Desired Features

  • Prompt Input

    • A text box where users can enter prompts for the LLM
  • Dropdown Selectors

    • Engine: Choose between TornadoVM, JVM, etc.
    • Model: Select the model file to use for inference between Llama, Mistral etc
  • File Picker

    • A "Browse" button to select .java files or model directories
  • Run Button

    • A button to trigger inference (e.g., using tornado-llama-opencv)
  • Output Display

    • A read-only area to show model responses and logs
    • Optional: System Monitoring Panel
    • Live GPU usage via nvtop or nvidia-smi
    • Live CPU & memory stats via htop or Java system metrics

Preferably, integrated with the core code base with JavaFX, or external with QT or python QT.

Example POC:

Image

mikepapadim avatar Jun 13 '25 10:06 mikepapadim

I'd like to take on this issue, but does this feature also depend on #17 being merged first?

svntax avatar Jun 18 '25 05:06 svntax

Hello @svntax, no this issue is completely indenpendent from other PRs.

mikepapadim avatar Jun 18 '25 05:06 mikepapadim