Dockerfile#

See here for the main Dockerfile to construct the image for running an OpenAI compatible server with vLLM. More information about deploying with Docker can be found here.

Below is a visual representation of the multi-stage Dockerfile. The build graph contains the following nodes:

  • All build stages

  • The default build target (highlighted in grey)

  • External images (with dashed borders)

The edges of the build graph represent:

  • FROM … dependencies (with a solid line and a full arrow head)

  • COPY –from=… dependencies (with a dashed line and an empty arrow head)

  • RUN –mount=(.*)from=… dependencies (with a dotted line and an empty diamond arrow head)

    query

    Made using: patrickhoefler/dockerfilegraph

    Commands to regenerate the build graph (make sure to run it from the `root` directory of the vLLM repository where the dockerfile is present):

    dockerfilegraph -o png --legend --dpi 200 --max-label-length 50 --filename Dockerfile
    

    or in case you want to run it directly with the docker image:

    docker run \
       --rm \
       --user "$(id -u):$(id -g)" \
       --workdir /workspace \
       --volume "$(pwd)":/workspace \
       ghcr.io/patrickhoefler/dockerfilegraph:alpine \
       --output png \
       --dpi 200 \
       --max-label-length 50 \
       --filename Dockerfile \
       --legend
    

    (To run it for a different file, you can pass in a different argument to the flag –filename.)