1) Message boards : Sixtrack Application : XTrack suite (Message 8054)
Posted 6 Apr 2023 by Yeti
Post:
They said that the first versions of xtrack are only for cpu.

Laurence on 06. Febr 2023 wrote:
The Xtrack application is a new version of Sixtrack which supports GPUs. It is currently work in progress. The current status is that the application has been deployed correctly in the server and seems to be running. We hope to continue testing within the next few weeks.


So, the question is: Is Xtrack-Suite the same as XTrack-Beam-Simulation
2) Message boards : Sixtrack Application : XTrack suite (Message 8051)
Posted 4 Apr 2023 by Yeti
Post:
On what kind of GPUs will XTRack-Suite run?

My client(s) are only asking for AMD/ATI-GPU-Work, but I use NVIDIA:

19887 lhcathome-dev 04-04-2023 12:19 Sending scheduler request: To fetch work.
19888 lhcathome-dev 04-04-2023 12:19 Requesting new tasks for CPU and AMD/ATI GPU
19889 lhcathome-dev 04-04-2023 12:19 Scheduler request completed: got 0 new tasks
19890 lhcathome-dev 04-04-2023 12:19 No tasks sent
19891 lhcathome-dev 04-04-2023 12:19 No tasks are available for Xtrack beam simulation
19892 lhcathome-dev 04-04-2023 12:19 Project requested delay of 61 seconds
3) Message boards : ATLAS Application : ATLAS vbox and native 3.01 (Message 8040)
Posted 23 Mar 2023 by Yeti
Post:
Yeti,
it was a scientist during the generating of the tasks. It was a file not for Boinc.
You find the answer from CP in -prod.

Yeah, I had already read it, but someone should tell a word about when it will be fixed, will WUs with this mistake get cancelled from sending out or whatever

I posted it here again because I didn't want to fokus the whole community on this point
4) Message boards : ATLAS Application : ATLAS vbox and native 3.01 (Message 8038)
Posted 23 Mar 2023 by Yeti
Post:
No, it's still the old version in prod. It must be just new batches of tasks with large files. I'll ask the submitters why it's like this now.

Have stopped one Threadripper 3995 overnight.
80 MBit/s from ISP and 1 GBit/s Network (including Squid) running at the limit, since those 1 GByte-Atlas downloads are active.
All Atlas-Tasks finishing with Hits-File so long.

David, any news on this ?

Meanwhile I have stopped all Atlas-Downloads, these 1,2 GB for each WU are too much, since yesterday evening I have downloaded 0,4 Terrabyte only from Atlas-Servers
5) Message boards : ATLAS Application : Huge EVNT Files (Message 8034)
Posted 22 Mar 2023 by Yeti
Post:
Just noticed those huge ATLAS EVNT files being downloaded from prod to different clients:
1,166,414,003 => 1.2 GB each
Same here, I got tons of it and my vDSL is overloaded:

6) Message boards : ATLAS Application : ATLAS vbox and native 3.01 (Message 8033)
Posted 22 Mar 2023 by Yeti
Post:
Seeing 1,08 GByte download in production.
Is this new Version transfered from -dev?
Now a second download on the same PC with 1,09GByte in production.
Application for Atlas in prod is the old one??
In Germany we say "Holland in Not".
My vDSL works at it's limit, but it seems to be overloaded with the new 1.2 GB Atlas-Tasks in Live.



Normally I have at worktime a limit for Atlas with 50 MB Download, I have opened it, but it is still not enough :-(
7) Message boards : ATLAS Application : ATLAS vbox and native 3.01 (Message 8008)
Posted 21 Mar 2023 by Yeti
Post:
with stabil checkpointing 2.000 events with 3.01 wouldn't be a problem for a lot of volunteers inclusive me
8) Message boards : ATLAS Application : ATLAS vbox and native 3.01 (Message 7985)
Posted 20 Mar 2023 by Yeti
Post:
Sorry, but I'm against these smaller tasks. I think it was a good balance with the 200 Events-WUs and vote for 500 events in future

My reasons:

Downloads for the nrw VB-Tasks will be bigger, even if they still contain only 200 events. Did I understand this right ?

The more in Download from 200 to 500 events for VB-Tasks will be small. Am I right ?

Shure the upload will be bigger, but I guess that will only be a problem for Magic

The new 200 Events WUs will more than double the downloads. Much more than necassary !

On my machines with modern CPUs, Ubuntu in a VM and a central squid the run-times differ between 2 and 2,5 hours for one 4-Core-WU, in future that will be 1 hours. Not really much, from my side of view to short.

Perhaps you can make it configurable in LHC-Preferences and the user can switch between 200 and 500 tasks, that could help.
9) Message boards : ATLAS Application : ATLAS vbox and native 3.01 (Message 7977)
Posted 16 Mar 2023 by Yeti
Post:
Top show no athena.py
Is this correct?
Yes, it should show a python task, with "Cores * 100%", so my 4-Core-WU shows the Python task with 400% CPU-Usage
10) Message boards : ATLAS Application : ATLAS vbox and native 3.01 (Message 7966)
Posted 15 Mar 2023 by Yeti
Post:
Meanwhile I have finished several of the 500-Events-WUs and they all look fine so far
11) Message boards : ATLAS Application : ATLAS vbox and native 3.01 (Message 7965)
Posted 15 Mar 2023 by Yeti
Post:
A long setup time would be expected for the first native task, since the CVMFS cache needs to be filled with the new software libraries. I'd be interested to see the timing for subsequent tasks on the same host. On my test host running native tasks (inside CERN, so ideal conditions) and a warm cache it takes around 5 mins to start crunching.

I can not confirm this expected behaviour.

First: Remember, I have set up a central Squid-Proxy that works fine.

You can check all my active clients, they are all running one DEV-Task after another. So, I would expect that the second or third Task would need less startup time, but they all need the same time.

Example: https://lhcathomedev.cern.ch/lhcathome-dev/results.php?hostid=4703

In the log-file I see still these warnings:

[2023-03-14 18:19:53] 2023-03-14 17:19:10,304 | INFO | [attempt=2/3] loading data from url=https://atlas-cric.cern.ch/cache/ddmendpoints.json
[2023-03-14 18:19:53] 2023-03-14 17:19:10,396 | WARNING | failed to load data from url=https://atlas-cric.cern.ch/cache/ddmendpoints.json, error: <urlopen error [SSL: CERTIFICATE_VERIFY_FAILED] certificate verify failed:
[2023-03-14 18:19:53] 2023-03-14 17:19:10,397 | INFO | will try again after 18s..
[2023-03-14 18:19:53] 2023-03-14 17:19:28,484 | INFO | [attempt=3/3] loading data from url=https://atlas-cric.cern.ch/cache/ddmendpoints.json
[2023-03-14 18:19:53] 2023-03-14 17:19:28,571 | WARNING | failed to load data from url=https://atlas-cric.cern.ch/cache/ddmendpoints.json, error: <urlopen error [SSL: CERTIFICATE_VERIFY_FAILED] certificate verify failed:
[2023-03-14 18:19:53] 2023-03-14 17:19:28,572 | WARNING | cache file=/var/lib/boinc_data/boinc-01/slots/2/agis_ddmendpoints.agis.ALL.json is not available: [Errno 2] No such file or directory: '/var/lib/boinc_data/boinc-0
[2023-03-14 18:19:53] 2023-03-14 17:19:28,606 | INFO | transferring file log.32413688._000229-57598-1678802947.job.log.tgz.1 from /var/lib/boinc_data/boinc-01/slots/2/PanDA_Pilot-5753961892/log.32413688._000229-57598-1
[2023-03-14 18:19:53] 2023-03-14 17:19:28,606 | INFO | executing command: /usr/bin/env mv /var/lib/boinc_data/boinc-01/slots/2/PanDA_Pilot-5753961892/log.32413688._000229-57598-1678802947.job.log.tgz.1 /var/lib/boinc_d
[2023-03-14 18:19:53] 2023-03-14 17:19:28,627 | INFO | Adding to output.list: log.32413688._000229-57598-1678802947.job.log.tgz.1 davs://dav.ndgf.org:443/atlas/disk/atlasdatadisk/rucio/valid1/20/76/log.32413688._000229
[2023-03-14 18:19:53] 2023-03-14 17:19:28,628 | INFO | executing command: ps aux -q 306808
[2023-03-14 18:19:53] 2023-03-14 17:19:28,656 | INFO | summary of transferred files:
[2023-03-14 18:19:53] 2023-03-14 17:19:28,657 | INFO | -- lfn=log.32413688._000229-57598-1678802947.job.log.tgz.1, status_code=0, status=transferred
[2023-03-14 18:19:53] 2023-03-14 17:19:28,657 | INFO | stage-out finished correctly
[2023-03-14 18:19:53] 2023-03-14 17:19:28,704 | INFO | finished stage-out for finished payload, adding job to finished_jobs queue
[2023-03-14 18:19:53] 2023-03-14 17:19:29,298 | INFO | job 5753961892 has state=finished
[2023-03-14 18:19:53] 2023-03-14 17:19:29,299 | INFO | preparing for final server update for job 5753961892 in state='finished'
[2023-03-14 18:19:53] 2023-03-14 17:19:29,299 | INFO | pilot will not update the server (heartbeat message will be written to file)
[2023-03-14 18:19:53] 2023-03-14 17:19:29,300 | INFO | job 5753961892 has finished - writing final server update
[2023-03-14 18:19:53] 2023-03-14 17:19:29,300 | WARNING | format EVNTtoHITS has no such key: dbData
[2023-03-14 18:19:53] 2023-03-14 17:19:29,301 | WARNING | format EVNTtoHITS has no such key: dbTime
[2023-03-14 18:19:53] 2023-03-14 17:19:29,302 | WARNING | wrong length of table data, x=[1678813630.0, 1678813691.0, 1678813752.0, 1678813813.0, 1678813874.0], y=[1051252.0, 1090314.0, 1911720.0, 2161488.0, 2321179.0] (mu
[2023-03-14 18:19:53] 2023-03-14 17:19:29,303 | INFO | total number of processed events: 5 (read)
12) Message boards : ATLAS Application : ATLAS vbox and native 3.01 (Message 7953)
Posted 14 Mar 2023 by Yeti
Post:
Sorry, but for me, it looks very unusual . These 2 tasks seemed to run endless, CPU-Time was way to low:



I'm running Ubuntu 22.04.x

Oh, I see, results have got uploaded now, both say "Hits file was produced successfull": Not shure, if this is really true

https://lhcathomedev.cern.ch/lhcathome-dev/result.php?resultid=3193657

https://lhcathomedev.cern.ch/lhcathome-dev/result.php?resultid=3193668
13) Message boards : ATLAS Application : ATLAS native 1.22 (Message 7746)
Posted 17 Aug 2022 by Yeti
Post:
Meanwhile I have got 18 WUs, for me it looks as if they all have run fine so far: https://lhcathomedev.cern.ch/lhcathome-dev/results.php?userid=250
14) Message boards : ATLAS Application : ATLAS native 1.22 (Message 7744)
Posted 17 Aug 2022 by Yeti
Post:
So far, I don't get any Atlas-WU.

BOINC-Client is 7.16.6, is this modern enough or do I need 7.20.x ?
15) Message boards : ATLAS Application : ATLAS native 1.22 (Message 7737)
Posted 16 Aug 2022 by Yeti
Post:
I have already Atlas-Native running and would be happy to help testing here with apptainer.

At the moment, there is singulary installed on my Ubuntu 20.04.4 LTS

Can you tell me please the exact instructions how I can install apptainer to the boxes?

Is Apptainer the same as CentOS..... ?

yeti
16) Message boards : ATLAS Application : ATLAS long simulation 1.01 (Message 7164)
Posted 25 Mar 2021 by Yeti
Post:
Long tasks are now available on LHC@Home (in a beta application): https://lhcathome.cern.ch/lhcathome/forum_thread.php?id=5625

When can Windows-Only-Users expect to run these ?

Maybe we can test it here ...
17) Message boards : News : ATLAS load tests (Message 5380)
Posted 10 Mar 2018 by Yeti
Post:
We have passed the checkmark of 10.000 WUs out in the field. A little bit more and we will reach the critical number
18) Message boards : News : ATLAS load tests (Message 5378)
Posted 10 Mar 2018 by Yeti
Post:
We are getting low on Atlas-WUs, at the moment there are 0 available.

For a loadtest of the new backend it would be good to have as much WUs as Volunteers are polling for
19) Message boards : ATLAS Application : ATLAS v0.50 and 0.51 (Message 5332)
Posted 16 Jan 2018 by Yeti
Post:
This WU doesn't startup, it is sitting for 6 hours at this point:



EDIT: https://lhcathomedev.cern.ch/lhcathome-dev/result.php?resultid=383071
20) Message boards : Number crunching : Can anyone explain the badges? (Message 5059)
Posted 3 Aug 2017 by Yeti
Post:
I seem to have an "ATLASYearsBronze-gluon.png" badge, but no idea where it came from, or what it means. I see that others have this or other badges. Searched forums, find nothing about badges. Home page doesn't mention badges. "My account" page shows the badge but no links to get info. AFAIK, I've never done any ATLAS tasks? Done plenty of others, would expect a CMS badge if anything, since that's what I did most of the work on.

https://lhcathome.cern.ch/lhcathome/badges

Oh - the badge (like the credits and everything else) is not being exported to the stats sites, which STILL show this project as being "offline", for months now.

Why should they do, this is "only" a development-project. If you are for credits you should run Live-Project at https://lhcathome.cern.ch/lhcathome/


Next 20


©2024 CERN