| Name | Theory_2922-4883451-492_1 |
| Workunit | 2711797 |
| Created | 6 Feb 2026, 20:16:26 UTC |
| Sent | 6 Feb 2026, 23:28:52 UTC |
| Report deadline | 16 Feb 2026, 23:28:52 UTC |
| Received | 17 Feb 2026, 12:33:46 UTC |
| Server state | Over |
| Outcome | Computation error |
| Client state | Aborted by user |
| Exit status | 203 (0x000000CB) EXIT_ABORTED_VIA_GUI |
| Computer ID | 5531 |
| Run time | 10 min 21 sec |
| CPU time | 10 min 53 sec |
| Priority | 0 |
| Validate state | Invalid |
| Credit | 0.00 |
| Device peak FLOPS | 6.78 GFLOPS |
| Application version | Theory Simulation v7.65 (docker) windows_x86_64 |
| Peak working set size | 13.38 MB |
| Peak swap size | 2.43 MB |
| Peak disk usage | 1.60 GB |
<core_client_version>8.2.8</core_client_version>
<![CDATA[
<message>
aborted by user</message>
<stderr_txt>
thomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1
program: podman
command output:
92.85% 352MB / 16.73GB
EOM
running docker command: ps --all -f "name=boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1"
program: podman
command output:
CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES
e0e0c95d3e81 localhost/boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492:latest /bin/sh -c ./entr... 10 days ago Up 7 minutes 80/tcp boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1
EOM
running docker command: stats --no-stream --format "{{.CPUPerc}} {{.MemUsage}}" boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1
program: podman
command output:
93.09% 352.4MB / 16.73GB
EOM
running docker command: ps --all -f "name=boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1"
program: podman
command output:
CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES
e0e0c95d3e81 localhost/boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492:latest /bin/sh -c ./entr... 10 days ago Up 7 minutes 80/tcp boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1
EOM
running docker command: stats --no-stream --format "{{.CPUPerc}} {{.MemUsage}}" boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1
program: podman
command output:
93.34% 353.3MB / 16.73GB
EOM
running docker command: ps --all -f "name=boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1"
program: podman
command output:
CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES
e0e0c95d3e81 localhost/boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492:latest /bin/sh -c ./entr... 10 days ago Up 7 minutes 80/tcp boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1
EOM
running docker command: stats --no-stream --format "{{.CPUPerc}} {{.MemUsage}}" boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1
program: podman
command output:
93.56% 353.7MB / 16.73GB
EOM
running docker command: ps --all -f "name=boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1"
program: podman
command output:
CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES
e0e0c95d3e81 localhost/boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492:latest /bin/sh -c ./entr... 10 days ago Up 8 minutes 80/tcp boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1
EOM
running docker command: stats --no-stream --format "{{.CPUPerc}} {{.MemUsage}}" boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1
program: podman
command output:
93.78% 354.7MB / 16.73GB
EOM
running docker command: ps --all -f "name=boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1"
program: podman
command output:
CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES
e0e0c95d3e81 localhost/boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492:latest /bin/sh -c ./entr... 10 days ago Up 8 minutes 80/tcp boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1
EOM
running docker command: stats --no-stream --format "{{.CPUPerc}} {{.MemUsage}}" boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1
program: podman
command output:
93.99% 355.3MB / 16.73GB
EOM
running docker command: ps --all -f "name=boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1"
program: podman
command output:
CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES
e0e0c95d3e81 localhost/boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492:latest /bin/sh -c ./entr... 10 days ago Up 8 minutes 80/tcp boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1
EOM
running docker command: stats --no-stream --format "{{.CPUPerc}} {{.MemUsage}}" boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1
program: podman
command output:
94.18% 354.9MB / 16.73GB
EOM
running docker command: ps --all -f "name=boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1"
program: podman
command output:
CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES
e0e0c95d3e81 localhost/boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492:latest /bin/sh -c ./entr... 10 days ago Up 8 minutes 80/tcp boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1
EOM
running docker command: stats --no-stream --format "{{.CPUPerc}} {{.MemUsage}}" boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1
program: podman
command output:
94.38% 356.5MB / 16.73GB
EOM
running docker command: ps --all -f "name=boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1"
program: podman
command output:
CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES
e0e0c95d3e81 localhost/boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492:latest /bin/sh -c ./entr... 10 days ago Up 8 minutes 80/tcp boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1
EOM
running docker command: stats --no-stream --format "{{.CPUPerc}} {{.MemUsage}}" boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1
program: podman
command output:
94.57% 357MB / 16.73GB
EOM
running docker command: ps --all -f "name=boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1"
program: podman
command output:
CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES
e0e0c95d3e81 localhost/boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492:latest /bin/sh -c ./entr... 10 days ago Up 9 minutes 80/tcp boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1
EOM
running docker command: stats --no-stream --format "{{.CPUPerc}} {{.MemUsage}}" boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1
program: podman
command output:
94.74% 356.4MB / 16.73GB
EOM
running docker command: ps --all -f "name=boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1"
program: podman
command output:
CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES
e0e0c95d3e81 localhost/boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492:latest /bin/sh -c ./entr... 10 days ago Up 9 minutes 80/tcp boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1
EOM
running docker command: stats --no-stream --format "{{.CPUPerc}} {{.MemUsage}}" boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1
program: podman
command output:
94.90% 357.2MB / 16.73GB
EOM
running docker command: ps --all -f "name=boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1"
program: podman
command output:
CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES
e0e0c95d3e81 localhost/boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492:latest /bin/sh -c ./entr... 10 days ago Up 9 minutes 80/tcp boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1
EOM
running docker command: stats --no-stream --format "{{.CPUPerc}} {{.MemUsage}}" boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1
program: podman
command output:
95.09% 358.9MB / 16.73GB
EOM
running docker command: ps --all -f "name=boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1"
program: podman
command output:
CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES
e0e0c95d3e81 localhost/boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492:latest /bin/sh -c ./entr... 10 days ago Up 9 minutes 80/tcp boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1
EOM
running docker command: stats --no-stream --format "{{.CPUPerc}} {{.MemUsage}}" boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1
program: podman
command output:
95.24% 358.6MB / 16.73GB
EOM
running docker command: ps --all -f "name=boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1"
program: podman
command output:
CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES
e0e0c95d3e81 localhost/boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492:latest /bin/sh -c ./entr... 10 days ago Up 9 minutes 80/tcp boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1
EOM
running docker command: stats --no-stream --format "{{.CPUPerc}} {{.MemUsage}}" boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1
program: podman
command output:
95.39% 359.1MB / 16.73GB
EOM
running docker command: ps --all -f "name=boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1"
program: podman
command output:
CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES
e0e0c95d3e81 localhost/boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492:latest /bin/sh -c ./entr... 10 days ago Up 9 minutes 80/tcp boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1
EOM
running docker command: stats --no-stream --format "{{.CPUPerc}} {{.MemUsage}}" boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1
program: podman
command output:
95.55% 360.6MB / 16.73GB
EOM
running docker command: ps --all -f "name=boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1"
program: podman
command output:
CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES
e0e0c95d3e81 localhost/boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492:latest /bin/sh -c ./entr... 10 days ago Up 10 minutes 80/tcp boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1
EOM
running docker command: stats --no-stream --format "{{.CPUPerc}} {{.MemUsage}}" boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1
program: podman
command output:
95.68% 360.5MB / 16.73GB
EOM
running docker command: ps --all -f "name=boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1"
program: podman
command output:
CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES
e0e0c95d3e81 localhost/boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492:latest /bin/sh -c ./entr... 10 days ago Up 10 minutes 80/tcp boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1
EOM
running docker command: stats --no-stream --format "{{.CPUPerc}} {{.MemUsage}}" boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1
program: podman
command output:
95.82% 362.3MB / 16.73GB
EOM
running docker command: ps --all -f "name=boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1"
program: podman
command output:
CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES
e0e0c95d3e81 localhost/boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492:latest /bin/sh -c ./entr... 10 days ago Up 10 minutes 80/tcp boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1
EOM
running docker command: stats --no-stream --format "{{.CPUPerc}} {{.MemUsage}}" boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1
program: podman
command output:
95.96% 361.8MB / 16.73GB
EOM
running docker command: ps --all -f "name=boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1"
program: podman
command output:
CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES
e0e0c95d3e81 localhost/boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492:latest /bin/sh -c ./entr... 10 days ago Up 10 minutes 80/tcp boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1
EOM
running docker command: stats --no-stream --format "{{.CPUPerc}} {{.MemUsage}}" boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1
program: podman
command output:
96.08% 361.5MB / 16.73GB
EOM
running docker command: ps --all -f "name=boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1"
program: podman
command output:
CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES
e0e0c95d3e81 localhost/boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492:latest /bin/sh -c ./entr... 10 days ago Up 10 minutes 80/tcp boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1
EOM
running docker command: stats --no-stream --format "{{.CPUPerc}} {{.MemUsage}}" boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1
program: podman
command output:
96.22% 362.6MB / 16.73GB
EOM
running docker command: ps --all -f "name=boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1"
program: podman
command output:
CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES
e0e0c95d3e81 localhost/boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492:latest /bin/sh -c ./entr... 10 days ago Up 10 minutes 80/tcp boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1
EOM
running docker command: stats --no-stream --format "{{.CPUPerc}} {{.MemUsage}}" boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1
program: podman
command output:
96.33% 363.4MB / 16.73GB
EOM
running docker command: ps --all -f "name=boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1"
program: podman
command output:
CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES
e0e0c95d3e81 localhost/boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492:latest /bin/sh -c ./entr... 10 days ago Up 11 minutes 80/tcp boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1
EOM
running docker command: stats --no-stream --format "{{.CPUPerc}} {{.MemUsage}}" boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1
program: podman
command output:
96.44% 362.8MB / 16.73GB
EOM
running docker command: ps --all -f "name=boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1"
program: podman
command output:
CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES
e0e0c95d3e81 localhost/boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492:latest /bin/sh -c ./entr... 10 days ago Up 11 minutes 80/tcp boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1
EOM
running docker command: stats --no-stream --format "{{.CPUPerc}} {{.MemUsage}}" boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1
program: podman
command output:
96.57% 364MB / 16.73GB
EOM
running docker command: ps --all -f "name=boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1"
program: podman
command output:
CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES
e0e0c95d3e81 localhost/boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492:latest /bin/sh -c ./entr... 10 days ago Up 11 minutes 80/tcp boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1
EOM
running docker command: stats --no-stream --format "{{.CPUPerc}} {{.MemUsage}}" boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1
program: podman
command output:
96.66% 364.9MB / 16.73GB
EOM
running docker command: ps --all -f "name=boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1"
program: podman
command output:
CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES
e0e0c95d3e81 localhost/boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492:latest /bin/sh -c ./entr... 10 days ago Up 11 minutes 80/tcp boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1
EOM
running docker command: stats --no-stream --format "{{.CPUPerc}} {{.MemUsage}}" boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1
program: podman
command output:
96.76% 364.4MB / 16.73GB
EOM
running docker command: ps --all -f "name=boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1"
program: podman
command output:
CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES
e0e0c95d3e81 localhost/boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492:latest /bin/sh -c ./entr... 10 days ago Up 11 minutes 80/tcp boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1
EOM
running docker command: stats --no-stream --format "{{.CPUPerc}} {{.MemUsage}}" boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1
program: podman
command output:
96.88% 365.9MB / 16.73GB
EOM
running docker command: ps --all -f "name=boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1"
program: podman
command output:
CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES
e0e0c95d3e81 localhost/boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492:latest /bin/sh -c ./entr... 10 days ago Up 12 minutes 80/tcp boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1
EOM
running docker command: stats --no-stream --format "{{.CPUPerc}} {{.MemUsage}}" boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1
program: podman
command output:
96.97% 365.1MB / 16.73GB
EOM
running docker command: ps --all -f "name=boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1"
program: podman
command output:
CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES
e0e0c95d3e81 localhost/boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492:latest /bin/sh -c ./entr... 10 days ago Up 12 minutes 80/tcp boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1
EOM
running docker command: stats --no-stream --format "{{.CPUPerc}} {{.MemUsage}}" boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1
program: podman
command output:
97.06% 365.8MB / 16.73GB
EOM
running docker command: ps --all -f "name=boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1"
program: podman
command output:
CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES
e0e0c95d3e81 localhost/boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492:latest /bin/sh -c ./entr... 10 days ago Up 12 minutes 80/tcp boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1
EOM
running docker command: stats --no-stream --format "{{.CPUPerc}} {{.MemUsage}}" boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1
program: podman
command output:
97.17% 366.9MB / 16.73GB
EOM
running docker command: ps --all -f "name=boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1"
program: podman
command output:
CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES
e0e0c95d3e81 localhost/boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492:latest /bin/sh -c ./entr... 10 days ago Up 12 minutes 80/tcp boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1
EOM
running docker command: stats --no-stream --format "{{.CPUPerc}} {{.MemUsage}}" boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1
program: podman
command output:
97.25% 366.9MB / 16.73GB
EOM
running docker command: ps --all -f "name=boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1"
program: podman
command output:
CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES
e0e0c95d3e81 localhost/boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492:latest /bin/sh -c ./entr... 10 days ago Up 12 minutes 80/tcp boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1
EOM
running docker command: stats --no-stream --format "{{.CPUPerc}} {{.MemUsage}}" boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1
program: podman
command output:
97.35% 368.1MB / 16.73GB
EOM
running docker command: ps --all -f "name=boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1"
program: podman
command output:
CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES
e0e0c95d3e81 localhost/boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492:latest /bin/sh -c ./entr... 10 days ago Up 12 minutes 80/tcp boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1
EOM
running docker command: stats --no-stream --format "{{.CPUPerc}} {{.MemUsage}}" boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1
program: podman
command output:
97.43% 368MB / 16.73GB
EOM
running docker command: ps --all -f "name=boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1"
program: podman
command output:
CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES
e0e0c95d3e81 localhost/boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492:latest /bin/sh -c ./entr... 10 days ago Up 13 minutes 80/tcp boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1
EOM
running docker command: stats --no-stream --format "{{.CPUPerc}} {{.MemUsage}}" boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1
program: podman
command output:
97.50% 368.8MB / 16.73GB
EOM
running docker command: ps --all -f "name=boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1"
program: podman
command output:
CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES
e0e0c95d3e81 localhost/boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492:latest /bin/sh -c ./entr... 10 days ago Up 13 minutes 80/tcp boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1
EOM
running docker command: stats --no-stream --format "{{.CPUPerc}} {{.MemUsage}}" boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1
program: podman
command output:
97.59% 368.7MB / 16.73GB
EOM
running docker command: ps --all -f "name=boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1"
program: podman
command output:
CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES
e0e0c95d3e81 localhost/boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492:latest /bin/sh -c ./entr... 10 days ago Up 13 minutes 80/tcp boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1
EOM
running docker command: stats --no-stream --format "{{.CPUPerc}} {{.MemUsage}}" boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1
program: podman
command output:
97.66% 370.2MB / 16.73GB
EOM
running docker command: ps --all -f "name=boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1"
program: podman
command output:
CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES
e0e0c95d3e81 localhost/boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492:latest /bin/sh -c ./entr... 10 days ago Up 13 minutes 80/tcp boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1
EOM
running docker command: stats --no-stream --format "{{.CPUPerc}} {{.MemUsage}}" boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1
program: podman
command output:
97.74% 369.9MB / 16.73GB
EOM
running docker command: ps --all -f "name=boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1"
program: podman
command output:
CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES
e0e0c95d3e81 localhost/boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492:latest /bin/sh -c ./entr... 10 days ago Up 13 minutes 80/tcp boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1
EOM
running docker command: stats --no-stream --format "{{.CPUPerc}} {{.MemUsage}}" boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1
program: podman
command output:
97.82% 372MB / 16.73GB
EOM
running docker command: ps --all -f "name=boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1"
program: podman
command output:
CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES
e0e0c95d3e81 localhost/boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492:latest /bin/sh -c ./entr... 10 days ago Up 13 minutes 80/tcp boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1
EOM
running docker command: stats --no-stream --format "{{.CPUPerc}} {{.MemUsage}}" boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1
program: podman
command output:
97.89% 370.8MB / 16.73GB
EOM
got abort request from client
running docker command: kill boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1
program: podman
command output:
boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1
EOM
running docker command: logs boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1
program: podman
command output:
Could not find a local HTTP proxy
CVMFS and Frontier will have to use DIRECT connections
This makes the application less efficient
It also puts higher load on the project servers
Setting up a local HTTP proxy is highly recommended
Advice can be found in the project forum
Using custom CVMFS.
Probing CVMFS repositories ...
Probing /cvmfs/alice.cern.ch... OK
Probing /cvmfs/cvmfs-config.cern.ch... OK
Probing /cvmfs/grid.cern.ch... OK
Probing /cvmfs/sft.cern.ch... OK
Excerpt from "cvmfs_config stat":
VERSION HOST PROXY
2.13.3.0 http://s1cern-cvmfs.openhtc.io DIRECT
******************************************************************
IMPORTANT HINT(S)!
******************************************************************
CVMFS server: http://s1cern-cvmfs.openhtc.io
CVMFS proxy: DIRECT
No local HTTP proxy found.
With this setup concurrently running containers can't share
a common CVMFS cache. A local HTTP proxy is therefore
highly recommended.
More info how to configure a local HTTP proxy:
https://lhcathome.cern.ch/lhcathome/forum_thread.php?id=5473
https://lhcathome.cern.ch/lhcathome/forum_thread.php?id=5474
******************************************************************
Environment HTTP proxy: not set
job: htmld=/var/www/lighttpd
job: unpack exitcode=0
Could not find a local HTTP proxy
CVMFS and Frontier will have to use DIRECT connections
This makes the application less efficient
It also puts higher load on the project servers
Setting up a local HTTP proxy is highly recommended
Advice can be found in the project forum
Using custom CVMFS.
Probing CVMFS repositories ...
Probing /cvmfs/alice.cern.ch... OK
Probing /cvmfs/cvmfs-config.cern.ch... OK
Probing /cvmfs/grid.cern.ch... OK
Probing /cvmfs/sft.cern.ch... OK
Excerpt from "cvmfs_config stat":
VERSION HOST PROXY
2.13.3.0 http://s1cern-cvmfs.openhtc.io DIRECT
******************************************************************
IMPORTANT HINT(S)!
******************************************************************
CVMFS server: http://s1cern-cvmfs.openhtc.io
CVMFS proxy: DIRECT
No local HTTP proxy found.
With this setup concurrently running containers can't share
a common CVMFS cache. A local HTTP proxy is therefore
highly recommended.
More info how to configure a local HTTP proxy:
https://lhcathome.cern.ch/lhcathome/forum_thread.php?id=5473
https://lhcathome.cern.ch/lhcathome/forum_thread.php?id=5474
******************************************************************
Environment HTTP proxy: not set
job: htmld=/var/www/lighttpd
job: unpack exitcode=0
Could not find a local HTTP proxy
CVMFS and Frontier will have to use DIRECT connections
This makes the application less efficient
It also puts higher load on the project servers
Setting up a local HTTP proxy is highly recommended
Advice can be found in the project forum
Using custom CVMFS.
Probing CVMFS repositories ...
Probing /cvmfs/alice.cern.ch... OK
Probing /cvmfs/cvmfs-config.cern.ch... OK
Probing /cvmfs/grid.cern.ch... OK
Probing /cvmfs/sft.cern.ch... OK
Excerpt from "cvmfs_config stat":
VERSION HOST PROXY
2.13.3.0 http://s1cern-cvmfs.openhtc.io DIRECT
******************************************************************
IMPORTANT HINT(S)!
******************************************************************
CVMFS server: http://s1cern-cvmfs.openhtc.io
CVMFS proxy: DIRECT
No local HTTP proxy found.
With this setup concurrently running containers can't share
a common CVMFS cache. A local HTTP proxy is therefore
highly recommended.
More info how to configure a local HTTP proxy:
https://lhcathome.cern.ch/lhcathome/forum_thread.php?id=5473
https://lhcathome.cern.ch/lhcathome/forum_thread.php?id=5474
******************************************************************
Environment HTTP proxy: not set
job: htmld=/var/www/lighttpd
job: unpack exitcode=0
Could not find a local HTTP proxy
CVMFS and Frontier will have to use DIRECT connections
This makes the application less efficient
It also puts higher load on the project servers
Setting up a local HTTP proxy is highly recommended
Advice can be found in the project forum
Using custom CVMFS.
Probing CVMFS repositories ...
Probing /cvmfs/alice.cern.ch... OK
Probing /cvmfs/cvmfs-config.cern.ch... OK
Probing /cvmfs/grid.cern.ch... OK
Probing /cvmfs/sft.cern.ch... OK
Excerpt from "cvmfs_config stat":
VERSION HOST PROXY
2.13.3.0 http://s1cern-cvmfs.openhtc.io DIRECT
******************************************************************
IMPORTANT HINT(S)!
******************************************************************
CVMFS server: http://s1cern-cvmfs.openhtc.io
CVMFS proxy: DIRECT
No local HTTP proxy found.
With this setup concurrently running containers can't share
a common CVMFS cache. A local HTTP proxy is therefore
highly recommended.
More info how to configure a local HTTP proxy:
https://lhcathome.cern.ch/lhcathome/forum_thread.php?id=5473
https://lhcathome.cern.ch/lhcathome/forum_thread.php?id=5474
******************************************************************
Environment HTTP proxy: not set
job: htmld=/var/www/lighttpd
job: unpack exitcode=0
Could not find a local HTTP proxy
CVMFS and Frontier will have to use DIRECT connections
This makes the application less efficient
It also puts higher load on the project servers
Setting up a local HTTP proxy is highly recommended
Advice can be found in the project forum
Using custom CVMFS.
Probing CVMFS repositories ...
Probing /cvmfs/alice.cern.ch... OK
Probing /cvmfs/cvmfs-config.cern.ch... OK
Probing /cvmfs/grid.cern.ch... OK
Probing /cvmfs/sft.cern.ch... OK
Excerpt from "cvmfs_config stat":
VERSION HOST PROXY
2.13.3.0 http://s1cern-cvmfs.openhtc.io DIRECT
******************************************************************
IMPORTANT HINT(S)!
******************************************************************
CVMFS server: http://s1cern-cvmfs.openhtc.io
CVMFS proxy: DIRECT
No local HTTP proxy found.
With this setup concurrently running containers can't share
a common CVMFS cache. A local HTTP proxy is therefore
highly recommended.
More info how to configure a local HTTP proxy:
https://lhcathome.cern.ch/lhcathome/forum_thread.php?id=5473
https://lhcathome.cern.ch/lhcathome/forum_thread.php?id=5474
******************************************************************
Environment HTTP proxy: not set
job: htmld=/var/www/lighttpd
job: unpack exitcode=0
Could not find a local HTTP proxy
CVMFS and Frontier will have to use DIRECT connections
This makes the application less efficient
It also puts higher load on the project servers
Setting up a local HTTP proxy is highly recommended
Advice can be found in the project forum
Using custom CVMFS.
Probing CVMFS repositories ...
Probing /cvmfs/alice.cern.ch... OK
Probing /cvmfs/cvmfs-config.cern.ch... OK
Probing /cvmfs/grid.cern.ch... OK
Probing /cvmfs/sft.cern.ch... OK
Excerpt from "cvmfs_config stat":
VERSION HOST PROXY
2.13.3.0 http://s1cern-cvmfs.openhtc.io DIRECT
******************************************************************
IMPORTANT HINT(S)!
******************************************************************
CVMFS server: http://s1cern-cvmfs.openhtc.io
CVMFS proxy: DIRECT
No local HTTP proxy found.
With this setup concurrently running containers can't share
a common CVMFS cache. A local HTTP proxy is therefore
highly recommended.
More info how to configure a local HTTP proxy:
https://lhcathome.cern.ch/lhcathome/forum_thread.php?id=5473
https://lhcathome.cern.ch/lhcathome/forum_thread.php?id=5474
******************************************************************
Environment HTTP proxy: not set
job: htmld=/var/www/lighttpd
job: unpack exitcode=0
Could not find a local HTTP proxy
CVMFS and Frontier will have to use DIRECT connections
This makes the application less efficient
It also puts higher load on the project servers
Setting up a local HTTP proxy is highly recommended
Advice can be found in the project forum
Using custom CVMFS.
Probing CVMFS repositories ...
Probing /cvmfs/alice.cern.ch... OK
Probing /cvmfs/cvmfs-config.cern.ch... OK
Probing /cvmfs/grid.cern.ch... OK
Probing /cvmfs/sft.cern.ch... OK
Excerpt from "cvmfs_config stat":
VERSION HOST PROXY
2.13.3.0 http://s1cern-cvmfs.openhtc.io DIRECT
******************************************************************
IMPORTANT HINT(S)!
******************************************************************
CVMFS server: http://s1cern-cvmfs.openhtc.io
CVMFS proxy: DIRECT
No local HTTP proxy found.
With this setup concurrently running containers can't share
a common CVMFS cache. A local HTTP proxy is therefore
highly recommended.
More info how to configure a local HTTP proxy:
https://lhcathome.cern.ch/lhcathome/forum_thread.php?id=5473
https://lhcathome.cern.ch/lhcathome/forum_thread.php?id=5474
******************************************************************
Environment HTTP proxy: not set
job: htmld=/var/www/lighttpd
job: unpack exitcode=0
Could not find a local HTTP proxy
CVMFS and Frontier will have to use DIRECT connections
This makes the application less efficient
It also puts higher load on the project servers
Setting up a local HTTP proxy is highly recommended
Advice can be found in the project forum
Using custom CVMFS.
Probing CVMFS repositories ...
Probing /cvmfs/alice.cern.ch... OK
Probing /cvmfs/cvmfs-config.cern.ch... OK
Probing /cvmfs/grid.cern.ch... OK
Probing /cvmfs/sft.cern.ch... OK
Excerpt from "cvmfs_config stat":
VERSION HOST PROXY
2.13.3.0 http://s1cern-cvmfs.openhtc.io DIRECT
******************************************************************
IMPORTANT HINT(S)!
******************************************************************
CVMFS server: http://s1cern-cvmfs.openhtc.io
CVMFS proxy: DIRECT
No local HTTP proxy found.
With this setup concurrently running containers can't share
a common CVMFS cache. A local HTTP proxy is therefore
highly recommended.
More info how to configure a local HTTP proxy:
https://lhcathome.cern.ch/lhcathome/forum_thread.php?id=5473
https://lhcathome.cern.ch/lhcathome/forum_thread.php?id=5474
******************************************************************
Environment HTTP proxy: not set
job: htmld=/var/www/lighttpd
job: unpack exitcode=0
Could not find a local HTTP proxy
CVMFS and Frontier will have to use DIRECT connections
This makes the application less efficient
It also puts higher load on the project servers
Setting up a local HTTP proxy is highly recommended
Advice can be found in the project forum
Using custom CVMFS.
Probing CVMFS repositories ...
Probing /cvmfs/alice.cern.ch... OK
Probing /cvmfs/cvmfs-config.cern.ch... OK
Probing /cvmfs/grid.cern.ch... OK
Probing /cvmfs/sft.cern.ch... OK
Excerpt from "cvmfs_config stat":
VERSION HOST PROXY
2.13.3.0 http://s1cern-cvmfs.openhtc.io DIRECT
******************************************************************
IMPORTANT HINT(S)!
******************************************************************
CVMFS server: http://s1cern-cvmfs.openhtc.io
CVMFS proxy: DIRECT
No local HTTP proxy found.
With this setup concurrently running containers can't share
a common CVMFS cache. A local HTTP proxy is therefore
highly recommended.
More info how to configure a local HTTP proxy:
https://lhcathome.cern.ch/lhcathome/forum_thread.php?id=5473
https://lhcathome.cern.ch/lhcathome/forum_thread.php?id=5474
******************************************************************
Environment HTTP proxy: not set
job: htmld=/var/www/lighttpd
job: unpack exitcode=0
Could not find a local HTTP proxy
CVMFS and Frontier will have to use DIRECT connections
This makes the application less efficient
It also puts higher load on the project servers
Setting up a local HTTP proxy is highly recommended
Advice can be found in the project forum
Using custom CVMFS.
Probing CVMFS repositories ...
Probing /cvmfs/alice.cern.ch... OK
Probing /cvmfs/cvmfs-config.cern.ch... OK
Probing /cvmfs/grid.cern.ch... OK
Probing /cvmfs/sft.cern.ch... OK
Excerpt from "cvmfs_config stat":
VERSION HOST PROXY
2.13.3.0 http://s1cern-cvmfs.openhtc.io DIRECT
******************************************************************
IMPORTANT HINT(S)!
******************************************************************
CVMFS server: http://s1cern-cvmfs.openhtc.io
CVMFS proxy: DIRECT
No local HTTP proxy found.
With this setup concurrently running containers can't share
a common CVMFS cache. A local HTTP proxy is therefore
highly recommended.
More info how to configure a local HTTP proxy:
https://lhcathome.cern.ch/lhcathome/forum_thread.php?id=5473
https://lhcathome.cern.ch/lhcathome/forum_thread.php?id=5474
******************************************************************
Environment HTTP proxy: not set
job: htmld=/var/www/lighttpd
job: unpack exitcode=0
Could not find a local HTTP proxy
CVMFS and Frontier will have to use DIRECT connections
This makes the application less efficient
It also puts higher load on the project servers
Setting up a local HTTP proxy is highly recommended
Advice can be found in the project forum
Using custom CVMFS.
Probing CVMFS repositories ...
Probing /cvmfs/alice.cern.ch... OK
Probing /cvmfs/cvmfs-config.cern.ch... OK
Probing /cvmfs/grid.cern.ch... OK
Probing /cvmfs/sft.cern.ch... OK
Excerpt from "cvmfs_config stat":
VERSION HOST PROXY
2.13.3.0 http://s1cern-cvmfs.openhtc.io DIRECT
******************************************************************
IMPORTANT HINT(S)!
******************************************************************
CVMFS server: http://s1cern-cvmfs.openhtc.io
CVMFS proxy: DIRECT
No local HTTP proxy found.
With this setup concurrently running containers can't share
a common CVMFS cache. A local HTTP proxy is therefore
highly recommended.
More info how to configure a local HTTP proxy:
https://lhcathome.cern.ch/lhcathome/forum_thread.php?id=5473
https://lhcathome.cern.ch/lhcathome/forum_thread.php?id=5474
******************************************************************
Environment HTTP proxy: not set
job: htmld=/var/www/lighttpd
job: unpack exitcode=0
Could not find a local HTTP proxy
CVMFS and Frontier will have to use DIRECT connections
This makes the application less efficient
It also puts higher load on the project servers
Setting up a local HTTP proxy is highly recommended
Advice can be found in the project forum
Using custom CVMFS.
Probing CVMFS repositories ...
Probing /cvmfs/alice.cern.ch... OK
Probing /cvmfs/cvmfs-config.cern.ch... OK
Probing /cvmfs/grid.cern.ch... OK
Probing /cvmfs/sft.cern.ch... OK
Excerpt from "cvmfs_config stat":
VERSION HOST PROXY
2.13.3.0 http://s1cern-cvmfs.openhtc.io DIRECT
******************************************************************
IMPORTANT HINT(S)!
******************************************************************
CVMFS server: http://s1cern-cvmfs.openhtc.io
CVMFS proxy: DIRECT
No local HTTP proxy found.
With this setup concurrently running containers can't share
a common CVMFS cache. A local HTTP proxy is therefore
highly recommended.
More info how to configure a local HTTP proxy:
https://lhcathome.cern.ch/lhcathome/forum_thread.php?id=5473
https://lhcathome.cern.ch/lhcathome/forum_thread.php?id=5474
******************************************************************
Environment HTTP proxy: not set
job: htmld=/var/www/lighttpd
job: unpack exitcode=0
Could not find a local HTTP proxy
CVMFS and Frontier will have to use DIRECT connections
This makes the application less efficient
It also puts higher load on the project servers
Setting up a local HTTP proxy is highly recommended
Advice can be found in the project forum
Using custom CVMFS.
Probing CVMFS repositories ...
Probing /cvmfs/alice.cern.ch... OK
Probing /cvmfs/cvmfs-config.cern.ch... OK
Probing /cvmfs/grid.cern.ch... OK
Probing /cvmfs/sft.cern.ch... OK
Excerpt from "cvmfs_config stat":
VERSION HOST PROXY
2.13.3.0 http://s1cern-cvmfs.openhtc.io DIRECT
******************************************************************
IMPORTANT HINT(S)!
******************************************************************
CVMFS server: http://s1cern-cvmfs.openhtc.io
CVMFS proxy: DIRECT
No local HTTP proxy found.
With this setup concurrently running containers can't share
a common CVMFS cache. A local HTTP proxy is therefore
highly recommended.
More info how to configure a local HTTP proxy:
https://lhcathome.cern.ch/lhcathome/forum_thread.php?id=5473
https://lhcathome.cern.ch/lhcathome/forum_thread.php?id=5474
******************************************************************
Environment HTTP proxy: not set
job: htmld=/var/www/lighttpd
job: unpack exitcode=0
EOM
stderr from container:
Could not find a local HTTP proxy
CVMFS and Frontier will have to use DIRECT connections
This makes the application less efficient
It also puts higher load on the project servers
Setting up a local HTTP proxy is highly recommended
Advice can be found in the project forum
Using custom CVMFS.
Probing CVMFS repositories ...
Probing /cvmfs/alice.cern.ch... OK
Probing /cvmfs/cvmfs-config.cern.ch... OK
Probing /cvmfs/grid.cern.ch... OK
Probing /cvmfs/sft.cern.ch... OK
Excerpt from "cvmfs_config stat":
VERSION HOST PROXY
2.13.3.0 http://s1cern-cvmfs.openhtc.io DIRECT
******************************************************************
IMPORTANT HINT(S)!
******************************************************************
CVMFS server: http://s1cern-cvmfs.openhtc.io
CVMFS proxy: DIRECT
No local HTTP proxy found.
With this setup concurrently running containers can't share
a common CVMFS cache. A local HTTP proxy is therefore
highly recommended.
More info how to configure a local HTTP proxy:
https://lhcathome.cern.ch/lhcathome/forum_thread.php?id=5473
https://lhcathome.cern.ch/lhcathome/forum_thread.php?id=5474
******************************************************************
Environment HTTP proxy: not set
job: htmld=/var/www/lighttpd
job: unpack exitcode=0
Could not find a local HTTP proxy
CVMFS and Frontier will have to use DIRECT connections
This makes the application less efficient
It also puts higher load on the project servers
Setting up a local HTTP proxy is highly recommended
Advice can be found in the project forum
Using custom CVMFS.
Probing CVMFS repositories ...
Probing /cvmfs/alice.cern.ch... OK
Probing /cvmfs/cvmfs-config.cern.ch... OK
Probing /cvmfs/grid.cern.ch... OK
Probing /cvmfs/sft.cern.ch... OK
Excerpt from "cvmfs_config stat":
VERSION HOST PROXY
2.13.3.0 http://s1cern-cvmfs.openhtc.io DIRECT
******************************************************************
IMPORTANT HINT(S)!
******************************************************************
CVMFS server: http://s1cern-cvmfs.openhtc.io
CVMFS proxy: DIRECT
No local HTTP proxy found.
With this setup concurrently running containers can't share
a common CVMFS cache. A local HTTP proxy is therefore
highly recommended.
More info how to configure a local HTTP proxy:
https://lhcathome.cern.ch/lhcathome/forum_thread.php?id=5473
https://lhcathome.cern.ch/lhcathome/forum_thread.php?id=5474
******************************************************************
Environment HTTP proxy: not set
job: htmld=/var/www/lighttpd
job: unpack exitcode=0
Could not find a local HTTP proxy
CVMFS and Frontier will have to use DIRECT connections
This makes the application less efficient
It also puts higher load on the project servers
Setting up a local HTTP proxy is highly recommended
Advice can be found in the project forum
Using custom CVMFS.
Probing CVMFS repositories ...
Probing /cvmfs/alice.cern.ch... OK
Probing /cvmfs/cvmfs-config.cern.ch... OK
Probing /cvmfs/grid.cern.ch... OK
Probing /cvmfs/sft.cern.ch... OK
Excerpt from "cvmfs_config stat":
VERSION HOST PROXY
2.13.3.0 http://s1cern-cvmfs.openhtc.io DIRECT
******************************************************************
IMPORTANT HINT(S)!
******************************************************************
CVMFS server: http://s1cern-cvmfs.openhtc.io
CVMFS proxy: DIRECT
No local HTTP proxy found.
With this setup concurrently running containers can't share
a common CVMFS cache. A local HTTP proxy is therefore
highly recommended.
More info how to configure a local HTTP proxy:
https://lhcathome.cern.ch/lhcathome/forum_thread.php?id=5473
https://lhcathome.cern.ch/lhcathome/forum_thread.php?id=5474
******************************************************************
Environment HTTP proxy: not set
job: htmld=/var/www/lighttpd
job: unpack exitcode=0
Could not find a local HTTP proxy
CVMFS and Frontier will have to use DIRECT connections
This makes the application less efficient
It also puts higher load on the project servers
Setting up a local HTTP proxy is highly recommended
Advice can be found in the project forum
Using custom CVMFS.
Probing CVMFS repositories ...
Probing /cvmfs/alice.cern.ch... OK
Probing /cvmfs/cvmfs-config.cern.ch... OK
Probing /cvmfs/grid.cern.ch... OK
Probing /cvmfs/sft.cern.ch... OK
Excerpt from "cvmfs_config stat":
VERSION HOST PROXY
2.13.3.0 http://s1cern-cvmfs.openhtc.io DIRECT
******************************************************************
IMPORTANT HINT(S)!
******************************************************************
CVMFS server: http://s1cern-cvmfs.openhtc.io
CVMFS proxy: DIRECT
No local HTTP proxy found.
With this setup concurrently running containers can't share
a common CVMFS cache. A local HTTP proxy is therefore
highly recommended.
More info how to configure a local HTTP proxy:
https://lhcathome.cern.ch/lhcathome/forum_thread.php?id=5473
https://lhcathome.cern.ch/lhcathome/forum_thread.php?id=5474
******************************************************************
Environment HTTP proxy: not set
job: htmld=/var/www/lighttpd
job: unpack exitcode=0
Could not find a local HTTP proxy
CVMFS and Frontier will have to use DIRECT connections
This makes the application less efficient
It also puts higher load on the project servers
Setting up a local HTTP proxy is highly recommended
Advice can be found in the project forum
Using custom CVMFS.
Probing CVMFS repositories ...
Probing /cvmfs/alice.cern.ch... OK
Probing /cvmfs/cvmfs-config.cern.ch... OK
Probing /cvmfs/grid.cern.ch... OK
Probing /cvmfs/sft.cern.ch... OK
Excerpt from "cvmfs_config stat":
VERSION HOST PROXY
2.13.3.0 http://s1cern-cvmfs.openhtc.io DIRECT
******************************************************************
IMPORTANT HINT(S)!
******************************************************************
CVMFS server: http://s1cern-cvmfs.openhtc.io
CVMFS proxy: DIRECT
No local HTTP proxy found.
With this setup concurrently running containers can't share
a common CVMFS cache. A local HTTP proxy is therefore
highly recommended.
More info how to configure a local HTTP proxy:
https://lhcathome.cern.ch/lhcathome/forum_thread.php?id=5473
https://lhcathome.cern.ch/lhcathome/forum_thread.php?id=5474
******************************************************************
Environment HTTP proxy: not set
job: htmld=/var/www/lighttpd
job: unpack exitcode=0
Could not find a local HTTP proxy
CVMFS and Frontier will have to use DIRECT connections
This makes the application less efficient
It also puts higher load on the project servers
Setting up a local HTTP proxy is highly recommended
Advice can be found in the project forum
Using custom CVMFS.
Probing CVMFS repositories ...
Probing /cvmfs/alice.cern.ch... OK
Probing /cvmfs/cvmfs-config.cern.ch... OK
Probing /cvmfs/grid.cern.ch... OK
Probing /cvmfs/sft.cern.ch... OK
Excerpt from "cvmfs_config stat":
VERSION HOST PROXY
2.13.3.0 http://s1cern-cvmfs.openhtc.io DIRECT
******************************************************************
IMPORTANT HINT(S)!
******************************************************************
CVMFS server: http://s1cern-cvmfs.openhtc.io
CVMFS proxy: DIRECT
No local HTTP proxy found.
With this setup concurrently running containers can't share
a common CVMFS cache. A local HTTP proxy is therefore
highly recommended.
More info how to configure a local HTTP proxy:
https://lhcathome.cern.ch/lhcathome/forum_thread.php?id=5473
https://lhcathome.cern.ch/lhcathome/forum_thread.php?id=5474
******************************************************************
Environment HTTP proxy: not set
job: htmld=/var/www/lighttpd
job: unpack exitcode=0
Could not find a local HTTP proxy
CVMFS and Frontier will have to use DIRECT connections
This makes the application less efficient
It also puts higher load on the project servers
Setting up a local HTTP proxy is highly recommended
Advice can be found in the project forum
Using custom CVMFS.
Probing CVMFS repositories ...
Probing /cvmfs/alice.cern.ch... OK
Probing /cvmfs/cvmfs-config.cern.ch... OK
Probing /cvmfs/grid.cern.ch... OK
Probing /cvmfs/sft.cern.ch... OK
Excerpt from "cvmfs_config stat":
VERSION HOST PROXY
2.13.3.0 http://s1cern-cvmfs.openhtc.io DIRECT
******************************************************************
IMPORTANT HINT(S)!
******************************************************************
CVMFS server: http://s1cern-cvmfs.openhtc.io
CVMFS proxy: DIRECT
No local HTTP proxy found.
With this setup concurrently running containers can't share
a common CVMFS cache. A local HTTP proxy is therefore
highly recommended.
More info how to configure a local HTTP proxy:
https://lhcathome.cern.ch/lhcathome/forum_thread.php?id=5473
https://lhcathome.cern.ch/lhcathome/forum_thread.php?id=5474
******************************************************************
Environment HTTP proxy: not set
job: htmld=/var/www/lighttpd
job: unpack exitcode=0
Could not find a local HTTP proxy
CVMFS and Frontier will have to use DIRECT connections
This makes the application less efficient
It also puts higher load on the project servers
Setting up a local HTTP proxy is highly recommended
Advice can be found in the project forum
Using custom CVMFS.
Probing CVMFS repositories ...
Probing /cvmfs/alice.cern.ch... OK
Probing /cvmfs/cvmfs-config.cern.ch... OK
Probing /cvmfs/grid.cern.ch... OK
Probing /cvmfs/sft.cern.ch... OK
Excerpt from "cvmfs_config stat":
VERSION HOST PROXY
2.13.3.0 http://s1cern-cvmfs.openhtc.io DIRECT
******************************************************************
IMPORTANT HINT(S)!
******************************************************************
CVMFS server: http://s1cern-cvmfs.openhtc.io
CVMFS proxy: DIRECT
No local HTTP proxy found.
With this setup concurrently running containers can't share
a common CVMFS cache. A local HTTP proxy is therefore
highly recommended.
More info how to configure a local HTTP proxy:
https://lhcathome.cern.ch/lhcathome/forum_thread.php?id=5473
https://lhcathome.cern.ch/lhcathome/forum_thread.php?id=5474
******************************************************************
Environment HTTP proxy: not set
job: htmld=/var/www/lighttpd
job: unpack exitcode=0
Could not find a local HTTP proxy
CVMFS and Frontier will have to use DIRECT connections
This makes the application less efficient
It also puts higher load on the project servers
Setting up a local HTTP proxy is highly recommended
Advice can be found in the project forum
Using custom CVMFS.
Probing CVMFS repositories ...
Probing /cvmfs/alice.cern.ch... OK
Probing /cvmfs/cvmfs-config.cern.ch... OK
Probing /cvmfs/grid.cern.ch... OK
Probing /cvmfs/sft.cern.ch... OK
Excerpt from "cvmfs_config stat":
VERSION HOST PROXY
2.13.3.0 http://s1cern-cvmfs.openhtc.io DIRECT
******************************************************************
IMPORTANT HINT(S)!
******************************************************************
CVMFS server: http://s1cern-cvmfs.openhtc.io
CVMFS proxy: DIRECT
No local HTTP proxy found.
With this setup concurrently running containers can't share
a common CVMFS cache. A local HTTP proxy is therefore
highly recommended.
More info how to configure a local HTTP proxy:
https://lhcathome.cern.ch/lhcathome/forum_thread.php?id=5473
https://lhcathome.cern.ch/lhcathome/forum_thread.php?id=5474
******************************************************************
Environment HTTP proxy: not set
job: htmld=/var/www/lighttpd
job: unpack exitcode=0
Could not find a local HTTP proxy
CVMFS and Frontier will have to use DIRECT connections
This makes the application less efficient
It also puts higher load on the project servers
Setting up a local HTTP proxy is highly recommended
Advice can be found in the project forum
Using custom CVMFS.
Probing CVMFS repositories ...
Probing /cvmfs/alice.cern.ch... OK
Probing /cvmfs/cvmfs-config.cern.ch... OK
Probing /cvmfs/grid.cern.ch... OK
Probing /cvmfs/sft.cern.ch... OK
Excerpt from "cvmfs_config stat":
VERSION HOST PROXY
2.13.3.0 http://s1cern-cvmfs.openhtc.io DIRECT
******************************************************************
IMPORTANT HINT(S)!
******************************************************************
CVMFS server: http://s1cern-cvmfs.openhtc.io
CVMFS proxy: DIRECT
No local HTTP proxy found.
With this setup concurrently running containers can't share
a common CVMFS cache. A local HTTP proxy is therefore
highly recommended.
More info how to configure a local HTTP proxy:
https://lhcathome.cern.ch/lhcathome/forum_thread.php?id=5473
https://lhcathome.cern.ch/lhcathome/forum_thread.php?id=5474
******************************************************************
Environment HTTP proxy: not set
job: htmld=/var/www/lighttpd
job: unpack exitcode=0
Could not find a local HTTP proxy
CVMFS and Frontier will have to use DIRECT connections
This makes the application less efficient
It also puts higher load on the project servers
Setting up a local HTTP proxy is highly recommended
Advice can be found in the project forum
Using custom CVMFS.
Probing CVMFS repositories ...
Probing /cvmfs/alice.cern.ch... OK
Probing /cvmfs/cvmfs-config.cern.ch... OK
Probing /cvmfs/grid.cern.ch... OK
Probing /cvmfs/sft.cern.ch... OK
Excerpt from "cvmfs_config stat":
VERSION HOST PROXY
2.13.3.0 http://s1cern-cvmfs.openhtc.io DIRECT
******************************************************************
IMPORTANT HINT(S)!
******************************************************************
CVMFS server: http://s1cern-cvmfs.openhtc.io
CVMFS proxy: DIRECT
No local HTTP proxy found.
With this setup concurrently running containers can't share
a common CVMFS cache. A local HTTP proxy is therefore
highly recommended.
More info how to configure a local HTTP proxy:
https://lhcathome.cern.ch/lhcathome/forum_thread.php?id=5473
https://lhcathome.cern.ch/lhcathome/forum_thread.php?id=5474
******************************************************************
Environment HTTP proxy: not set
job: htmld=/var/www/lighttpd
job: unpack exitcode=0
Could not find a local HTTP proxy
CVMFS and Frontier will have to use DIRECT connections
This makes the application less efficient
It also puts higher load on the project servers
Setting up a local HTTP proxy is highly recommended
Advice can be found in the project forum
Using custom CVMFS.
Probing CVMFS repositories ...
Probing /cvmfs/alice.cern.ch... OK
Probing /cvmfs/cvmfs-config.cern.ch... OK
Probing /cvmfs/grid.cern.ch... OK
Probing /cvmfs/sft.cern.ch... OK
Excerpt from "cvmfs_config stat":
VERSION HOST PROXY
2.13.3.0 http://s1cern-cvmfs.openhtc.io DIRECT
******************************************************************
IMPORTANT HINT(S)!
******************************************************************
CVMFS server: http://s1cern-cvmfs.openhtc.io
CVMFS proxy: DIRECT
No local HTTP proxy found.
With this setup concurrently running containers can't share
a common CVMFS cache. A local HTTP proxy is therefore
highly recommended.
More info how to configure a local HTTP proxy:
https://lhcathome.cern.ch/lhcathome/forum_thread.php?id=5473
https://lhcathome.cern.ch/lhcathome/forum_thread.php?id=5474
******************************************************************
Environment HTTP proxy: not set
job: htmld=/var/www/lighttpd
job: unpack exitcode=0
Could not find a local HTTP proxy
CVMFS and Frontier will have to use DIRECT connections
This makes the application less efficient
It also puts higher load on the project servers
Setting up a local HTTP proxy is highly recommended
Advice can be found in the project forum
Using custom CVMFS.
Probing CVMFS repositories ...
Probing /cvmfs/alice.cern.ch... OK
Probing /cvmfs/cvmfs-config.cern.ch... OK
Probing /cvmfs/grid.cern.ch... OK
Probing /cvmfs/sft.cern.ch... OK
Excerpt from "cvmfs_config stat":
VERSION HOST PROXY
2.13.3.0 http://s1cern-cvmfs.openhtc.io DIRECT
******************************************************************
IMPORTANT HINT(S)!
******************************************************************
CVMFS server: http://s1cern-cvmfs.openhtc.io
CVMFS proxy: DIRECT
No local HTTP proxy found.
With this setup concurrently running containers can't share
a common CVMFS cache. A local HTTP proxy is therefore
highly recommended.
More info how to configure a local HTTP proxy:
https://lhcathome.cern.ch/lhcathome/forum_thread.php?id=5473
https://lhcathome.cern.ch/lhcathome/forum_thread.php?id=5474
******************************************************************
Environment HTTP proxy: not set
job: htmld=/var/www/lighttpd
job: unpack exitcode=0
EOM
stderr end
running docker command: container rm boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1
program: podman
command output:
boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492_1
EOM
running docker command: image rm boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492
program: podman
command output:
Untagged: localhost/boinc__lhcathomedev.cern.ch_lhcathome-dev__theory_2922-4883451-492:latest
EOM
</stderr_txt>
]]>
©2026 CERN