testcontainers-python icon indicating copy to clipboard operation
testcontainers-python copied to clipboard

Feat : vLLM - adds object , adds tests , container , init , dataclasses

Open Josephrp opened this issue 2 months ago • 4 comments

solves https://github.com/testcontainers/testcontainers-python/issues/885

includes complete vLLM api , can be used for generation, embeddings , finetuning , dataset creation , testing experiment configurations , rl , and testing performance , ... etc.

Josephrp avatar Sep 30 '25 21:09 Josephrp

take a look at the example of a database container or the ollama container. they are intentionally minimal, giving you only what you need to get started. the purpose of this library is to provide temporary test infrastructure, not to implement every feature. if you can get this to the same amount of lines as ollama or less i may consider merging. looks like a promising project. if it implements the same api as ollama, maybe even lets consider making one container work for both images. not sure if this is possible/feasible.

alexanderankin avatar Oct 01 '25 13:10 alexanderankin

i see why you're saying this , and i understand where you're coming from , however ... it will be easier for me to show you the necessity & use of a fully featured object with a demo , so i'll do that in the next two weeks as i test , use and build with testcontainers . Meanwhile i'll dig into your comments and see if i can get it to where you're saying it should be :-)

Josephrp avatar Oct 03 '25 10:10 Josephrp

@Josephrp We are currently working on adding vLLM support into Docker Model Runner. So I think a much better approach for the project is adding integration with Docker Model Runner (similar to what Java and Go already did), which should also help with giving GPU access on e.g. macOS hosts (since we run inference natively).

https://java.testcontainers.org/modules/docker_model_runner/

kiview avatar Oct 06 '25 12:10 kiview

Thanks for highlighting this @kiview . Yes we are actively working on this, it is not ready, but please feel free to collaborate with us on this:

https://github.com/docker/inference-engine-vllm

https://github.com/vllm-project/vllm/pull/26160

ericcurtin avatar Oct 06 '25 13:10 ericcurtin