I see thanks - in my case the docker image seemed to lead to some kind of memory overflow when loading the models outside the runner, and the system froze. I did put them back into the runner.
I could not yet solve the problem that the container does not do inference, even if the bento does it with no problems. It seems the nvidia drivers are not installed in the container. Is this possible even if it works in the bento ?