Reuse inherited image's CMD or ENTRYPOINT

How can I include my own shell script CMD on container start/restart/attach, without removing the CMD used by an inherited image?

I am using this, which does execute my script fine, but appears to overwrite the PHP CMD:

  • How to connect locally hosted MySQL databse with the docker container
  • Docker doesn't download images (connection problems to registry)
  • What command can I use to find out the version of docker swarm I am using?
  • install mongoose in docker container
  • How are docker layer directories named?
  • Docker container can not run as a daemon
  • FROM php
    COPY /usr/local/bin
    CMD ["/usr/local/bin/"]

    What should I do differently? I am avoiding the prospect of copy/pasting the ENTRYPOINT or CMD of the parent image, and maybe that’s not a good approach.

  • Is it possible to install a bootable linux installer's payload into a docker container?
  • Nginx multiple load balancers or single load blanacer [closed]
  • mongodb replica set master “stateStr” : “REMOVED”
  • How to divert traffic from a docker container's ethernet interface to a bridge?
  • how to restrict number of service replicas in docker to 1
  • Docker push to private registry issues
  • 2 Solutions collect form web for “Reuse inherited image's CMD or ENTRYPOINT”

    As mentioned in the comments, there’s no built-in solution to this. From the Dockerfile, you can’t see the value of the current CMD or ENTRYPOINT. Having a run-parts solution is nice if you control the upstream base image and include this code there, allowing downstream components to make their changes. But docker there’s one inherent issue that will cause problems with this, containers should only run a single command that needs to run in the foreground. So if the upstream image kicks off, it would stay running without giving your later steps a chance to run, so you’re left with complexities to determine the order to run commands to ensure that a single command does eventually run without exiting.

    My personal preference is a much simpler and hardcoded option, to add my own command or entrypoint, and make the last step of my command to exec the upstream command. You will still need to manually identify the script name to call from the upstream Dockerfile. But now in your, you would have:

    # run various pieces of initialization code here
    # ...
    # kick off the upstream command:
    exec / "$@"

    By using an exec call, you transfer pid 1 to the upstream entrypoint so that signals get handled correctly. And the trailing "$@" passes through any command line arguments. You can use set to adjust the value of $@ if there are some args you want to process and extract in your own script.

    If the base image is not yours, you unfortunately have to call the parent command manually.

    If you own the parent image, you can try what the people at camptocamp suggest here.

    They basically use a generic script as an entry point that calls run-parts on a directory. What that does is run all scripts in that directory in lexicographic order. So when you extend an image, you just have to put your new scripts in that same folder.

    However, that means you’ll have to maintain order by prefixing your scripts which could potentially get out of hand. (Imagine the parent image decides to add a new script later…).

    Anyway, that could work.

    Update #1

    There is a long discussion on this docker compose issue about provisioning after container run. One suggestion is to wrap you docker run or compose command in a shell script and then run docker exec on your other commands.

    If you’d like to use that approach, you basically keep the parent CMD as the run command and you place yours as a docker exec after your docker run.

    Docker will be the best open platform for developers and sysadmins to build, ship, and run distributed applications.