We provide a docker/Dockerfile to construct the image for running an OpenAI compatible server with vLLM.
More information about deploying with Docker can be found here.
Below is a visual representation of the multi-stage Dockerfile. The build graph contains the following nodes:
All build stages
The default build target (highlighted in grey)
External images (with dashed borders)
The edges of the build graph represent:
FROM ... dependencies (with a solid line and a full arrow head)
COPY --from=... dependencies (with a dashed line and an empty arrow head)
RUN --mount=(.\*)from=... dependencies (with a dotted line and an empty diamond arrow head)