Posts by hadron

1) Message boards : Number crunching : Problems and Technical Issues with Rosetta@home (Message 112262)
Posted 19 Mar 2025 by hadron
Post:
Things are still very much broken- the amount of work In progress continues to fall, as does the number of Tasks processed per 24hrs.
Heaps of work available, but no one (or pretty much no one) can get any of it.

I'm not disagreeing, but where are you seeing that?
I've just looked at Rosetta project credits on Boincstats over the last 40 days and each 7 days over the last few weeks credit 195m, 207m, 282m, 282m, 277m
Or do you mean the last few days specifically?
Still very few unsent on the server status page - just 188 - but now 10.8m queued on the front page
I'm cutting off SiDock and WCG completely now - NNT - and giving Rosetta 100%

Check the home page. As of 1900 UTC today, there are over 10 million queued tasks.
2) Message boards : News : Outage notice (Message 111890)
Posted 10 Jan 2025 by hadron
Post:
Hi,
I think the whole Boinc system is down and has been so since 2024/12/07....

Well, it's not. LHC and Einstein are just fine, though LHC is in its winter break right now, so not much is happening there.
3) Message boards : Number crunching : Problems and Technical Issues with Rosetta@home (Message 109687)
Posted 28 Aug 2024 by hadron
Post:
My immediate reaction is to ask just what is the absolute minimum required to prevent Boinc from getting Rosetta beta tasks.
Either the project adds the option in the users account, or they support the anonymous platform function, or BOINC make changes to the supported values for the max_concurrent option in app_config.xml to make it possible to limit the number to less than one (ie- 0),
Any one of those 3 would make it possible.

I was actually asking if there was some way to do it with the way things are now.
4) Message boards : Number crunching : Problems and Technical Issues with Rosetta@home (Message 109682)
Posted 28 Aug 2024 by hadron
Post:
Respectfully disagree. Take LHC, for example.
Suppose you are running ATLAS and CMS only, both are mentioned in your app_config.xml
Stop right there & read what you posted, and then read what i posted.

As kotenok2000 pointed out, i was talking about using app_info.xml to make use of the Anonymous Platform function.
app_config.xml is completely, totally and utterly different.

TIL.... this is the first time that I can recall anyone mentioning "anonymous platform".
My immediate reaction is to ask just what is the absolute minimum required to prevent Boinc from getting Rosetta beta tasks.
5) Message boards : Number crunching : Problems and Technical Issues with Rosetta@home (Message 109679)
Posted 27 Aug 2024 by hadron
Post:
I am certain of this. There is no way in any file you can create or modify on your system that will determine whether or not you will be sent tasks for any particular app.
You can if the project supports anonymous platform.
The app_info.xml file contains the information for the applications that your system has, so you will receive work only for those applications & no others.

Respectfully disagree. Take LHC, for example.
Suppose you are running ATLAS and CMS only, both are mentioned in your app_config.xml file for whatever reasons (number of CPUs per task, max number of tasks for that app, for example). You have never run THEORY tasks, and they are de-selected in your user preferences on LHC. Your client knows nothing about them.
If you now decide you wish to start running THEORY tasks, you need only enable them in your preferences on LHC; no other action is necessary. The next time your client calls in requesting new tasks, you will get them -- guaranteed.
6) Message boards : Number crunching : Problems and Technical Issues with Rosetta@home (Message 109676)
Posted 27 Aug 2024 by hadron
Post:
This won't block Boinc from downloading beta work, and it has no effect all all on the number of cores each task will use.. It will only limit it from running more than X Rosetta tasks simultaneously.
If there were any way to tell Boinc not to download beta tasks, it would be in the cc_config.xml file. I see nothing in the user guide to suggest that is even possible. Therefore, since Rosetta itself does not distinguish between the "old' and beta tasks, it seems there is no way to block a client Boinc from downloading Rosetta beta tasks.

Are you sure? I used the xml file years ago and i remember it "reset" the project folder and started to download only the selected app.
But, probably, I remember wrong.

And, anyway, this is only attempts, until they decide to insert the possibility to choose in the user's profile...

I am certain of this. There is no way in any file you can create or modify on your system that will determine whether or not you will be sent tasks for any particular app. That can only be done on the project website, and Rosetta does not give you any choice.
7) Message boards : Number crunching : Problems and Technical Issues with Rosetta@home (Message 109642)
Posted 20 Aug 2024 by hadron
Post:
<max_concurrent>-1</max_concurrent> should have been unlimited.

Alternatively, don't even include the section at all, if you don't wish to limit the number of concurrent tasks.
8) Message boards : Number crunching : Problems and Technical Issues with Rosetta@home (Message 109640)
Posted 20 Aug 2024 by hadron
Post:
If you want to complain about how Boinc works, then you owe it to everyone to suggest an alternative.
So once again, how would you do it?
9) Message boards : Number crunching : Problems and Technical Issues with Rosetta@home (Message 109638)
Posted 20 Aug 2024 by hadron
Post:
Stuprd boinc thinks <max_concurrent>0</max_concurrent> is unlimited.

How would you do it?
10) Message boards : Number crunching : Problems and Technical Issues with Rosetta@home (Message 109637)
Posted 20 Aug 2024 by hadron
Post:
You can ask Veneto how to do that modification to block beta work. But I only got one beta and since then nothing but clean 4.20. So is it worth the effort to mess around?


The "usual" app_config.xml
Something like this (if i remember correctly):
<app_config>
    <app>
        <name>rosetta</name>
            <max_concurrent>X</max_concurrent>
    </app>
 </app_config>

(with X you can configure how many cores to use)

This won't block Boinc from downloading beta work, and it has no effect all all on the number of cores each task will use.. It will only limit it from running more than X Rosetta tasks simultaneously.

To control the number of cores a task uses, you need to include an <app_version> section in app_config.xml, like this:
<app_version>
        <app_name>ATLAS</app_name>
        <avg_ncpus>2</avg_ncpus>
        <plan_class>vbox64_mt_mcore_atlas</plan_class>
        <cmdline>--nthreads 2</cmdline>
    </app_version>

<avg_ncpus> tells Boinc that tasks will be running on 2 threads, while the cmdline parameter is passed to the program to tell it to do so.

If there were any way to tell Boinc not to download beta tasks, it would be in the cc_config.xml file. I see nothing in the user guide to suggest that is even possible. Therefore, since Rosetta itself does not distinguish between the "old' and beta tasks, it seems there is no way to block a client Boinc from downloading Rosetta beta tasks.
11) Message boards : Number crunching : Beyond newbie Q&A (Message 109560)
Posted 13 Aug 2024 by hadron
Post:
there's a report button at the bottom of each post.
Which does nothing.
When the spam first started i reported them as they occurred, but nothing was done about them. Apparently the project is happy to have the forums full of spam.

If a lot of people start reporting the posts, maybe the admins will be motivated to do something about it :)
12) Message boards : Number crunching : Beyond newbie Q&A (Message 109546)
Posted 11 Aug 2024 by hadron
Post:
Dedicated nursing writers ensure quality papers. They focus on delivering well-researched and accurate content. Each assignment is tailored to meet your specific needs. These writers have extensive knowledge in nursing. They understand academic standards and expectations. online class help services their goal is to help you succeed academically. With their expertise, you receive clear and concise papers. Every paper is original and free from plagiarism. Trust our dedicated writers for your nursing assignments. Achieve academic excellence with their support.



Oh look everyone..another spammer.
Go away Kinny unless you have something relevant to talk about.
We have enough spam on here.


there's a report button at the bottom of each post. Use that instead of posting a reply that will never be read by the intended recipient.
A reply to spam is just more spam.
13) Message boards : Number crunching : Problems and Technical Issues with Rosetta@home (Message 109511)
Posted 2 Aug 2024 by hadron
Post:
Haven't received any work units for two or three weeks now , what's up with that?

You haven't found the Boinc log file, or what?
If you would read that file, you would find this:

Fri 02 Aug 2024 04:53:12 AM | Rosetta@home | Scheduler request completed: got 0 new tasks
Fri 02 Aug 2024 04:53:12 AM | Rosetta@home | Project is temporarily shut down for maintenance

It will be back when it's back.
14) Message boards : Number crunching : Finally, a gpu app on Rosetta@Home (Message 109325)
Posted 31 May 2024 by hadron
Post:
Looks like a new CPU application, with, maybe, possibly, a GPU application at some stage (maybe...).


Yes.
But it's the first time that R@H admins speaks explicitly about work for gpu on our computers

On Windows, with an NVidia card -- whoopee. Call me when they come up with something that works.
15) Questions and Answers : Preferences : Rosetta uses only 1 core (Message 109324)
Posted 31 May 2024 by hadron
Post:
If I do get that GPU now, I might just dispense with running CPU-only tasks from Einstein, as the gravitational wave tasks seem to be taking 3/4 of an hour to complete, and earn 10K credits each!
A good GPU application can produce a huge amount of work.

My very first video card was a Hercules Graphics Card Plus with 64 kB of memory onboard. How far things have come since then!
Of course, it was mounted in a 1st generation 8086 clone running DOS 1.0, cost me $2000 -- no HDD, the driver for them didn't come out until DOS 1.1. What can you get for $680 these days? My first HDD cost that much -- a 20 MB Seagate. And then there was the printer -- a wide-carriage Epson dot matrix for around $1400.
So that's over $4000 -- what can you get for that these days?
I for one do not dream about a return to the "good old days" Kids these days do not have any clue how good they have it.
16) Questions and Answers : Preferences : Rosetta uses only 1 core (Message 109320)
Posted 30 May 2024 by hadron
Post:
And for the very last time -- the max_concurrent settings should not ever have any influence over Boinc requesting enough work for each project to be able to meet the resource share targets. The only effect they can possibly have is in how long it takes to reach those targets -- which I have painfully tried to suggest, I simply do not care.
I agree- limiting the number of Tasks that can run for a particular application won't stop your Resource Share settings from being met, it just increases the amount of time till it can be done (and that increase can be significant- with low core/thread systems with lots of active project at can take months. But even with your limits, the total number of cores/threads you've got available with only a few projects i would expect a week or two at most for your system (assuming no project issues- then all bets are off)).

And until you explicitly stated you didn't care how long it took, i had no idea that was your position. From the posts you were making i was under the impression you wanted things to sort themselves out sooner rather than later, hence my suggestion to remove the limits.


My apologies. I thought I was making my case clear -- evidently not.
Things are running fairly smoothly right now, but alas, no gamma ray pulsar tasks are currently available from Einstein, so I'm basically just running in wait-and-see mode until they return. In the interim, I've re-enabled the binary pulsar tasks (allowing 2 of them concurrently) just to keep the project's presence in Boinc's calculations. I might just bite the bullet now and get a second GPU to run the gravitational wave tasks. I have my eye on this one: https://www.newegg.ca/p/pl?N=100007708%20601110192%20601292089%20600100181%2050001315 but I need to check on Einstein to find out how many tasks I can run on it.

PS,Boinc seems to be working hard to overcome the massive beating my RAC on LHC took over the past few days -- 21 threads are allotted to LHC as I type.
Or just let things be for a while see how things end up.
You may have misunderstood what I was saying there -- apologies if I caused any confusion. Those 21 threads were all Boinc's doing, not mine.
Yep, i understood it as BOINC's doing, that's why i suggested there leaving things as they are instead of changing them further as i had suggested earlier in the post.

OK, just wanted to be sure.
Things may have settled down already. ATM, there are 11 Rosetta tasks running, along with 4 Atlas (2 threads each) and 1 Theory from LHC, and the 2 "placeholder" pulsar tasks from Einstein. If I do get that GPU now, I might just dispense with running CPU-only tasks from Einstein, as the gravitational wave tasks seem to be taking 3/4 of an hour to complete, and earn 10K credits each!
17) Questions and Answers : Preferences : Rosetta uses only 1 core (Message 109309)
Posted 29 May 2024 by hadron
Post:
Thus those first 3 lines actually represent the potential use of, not 12, but 20 threads.
Ok, but so what?
If BOINC needs them to honour your Resource Share it can make use of them, if not, then the other projects will make use of them.
But by putting in limits on the number of Tasks of a particular type that can be run, you hamstring BOINC's ability to get work in order to meet your Resource Share settings.

I was addressing this point you made -- I see it was in what I trimmed out:
As things are, you've told it what you want to happen with your Resource Share values, but then pretty much tied both it's arms behind it's back by limiting how much it is able to do (even though it's capable of doing much more), and it's just doing what it can, within those very restrictive limits.

You keep going back to the total amount of work being done here, suggesting that at times I will have periods during which there are several unused CPU threads.
I assure you, that has never been the case, and pretty much the only time it could possibly have happened is if LHC suddenly stopped delivering any tasks at all -- in which case, I could simply increase the limit on the other two projects until LHC returned to life.
Rather, the only problem I was having Boinc not requesting new LHC tasks, while it continued to grab all the Einstein stuff it could.
You keep talking about Resource Share settings, but have never asked me just what those settings were. Except for that very brief period when I set them to 400/200/100 (which I did mention), they were always the same for all three projects -- and, as I mentioned, they are all the same once more.
So I am having a hard time understanding why this is such an issue for you; regardless of any max_cpu settings, Boinc should always have been fetching work from all three projects, I would think in more or less in equal proportions.
And for the very last time -- the max_concurrent settings should not ever have any influence over Boinc requesting enough work for each project to be able to meet the resource share targets. The only effect they can possibly have is in how long it takes to reach those targets -- which I have painfully tried to suggest, I simply do not care.

Prepare for an avalanche of event log output.

Yeah, I already knew this -- I'll pass, thank you. It's all probably of use to no one but the folks who wrote the code.

Regardless of what the manual might say, personally, several times over the years, i have removed the the app_config file, exited & restarted BOINC with no issues. If there is no app_config file, then there are no values the Manager can read & then use. So it just runs with the applications defaults (And then later i've put in a modified app_info and just read it in for the values to take effect. I never tried it with the value set to 0, exiting BOINC and restarting after removing the file did the job of undoing what it had done).

Well, I figure it must be meaningful in some way, otherwise the warning would not be there in the first plce
Besides, it's a helluva lot easier just to replace a non-zero number with 0, save the file, then tell Boinc to read it in :D


PS,Boinc seems to be working hard to overcome the massive beating my RAC on LHC took over the past few days -- 21 threads are allotted to LHC as I type.
Or just let things be for a while see how things end up.

You may have misunderstood what I was saying there -- apologies if I caused any confusion. Those 21 threads were all Boinc's doing, not mine.
18) Questions and Answers : Preferences : Rosetta uses only 1 core (Message 109306)
Posted 28 May 2024 by hadron
Post:
Currently running tasks are: Rosetta and Einstein 8 each (controlled by app_config files)
In what way are the being controller by app_config files?
Whatever settings you've got there will impact on work Scheduling, as the Scheduler will have to work around those app_config settings.
LHC;
<max_concurrent>2</max_concurrent>
<max_concurrent>8</max_concurrent>
<max_concurrent>2</max_concurrent>
<project_max_concurrent>8</project_max_concurrent>
<max_concurrent>8</max_concurrent>
<max_concurrent>2</max_concurrent>
<max_concurrent>4</max_concurrent>
Ok, there's what's limiting how much work & of what type your system can do.

OK, let's take this in stages.
First off those last 2 lines represent Einstein sub-projects that I will never run again. They are still present in the app_config file only because I haven't got around to removing them -- to do so would require a complete project reset.
Next, the first and third lines are for CMS and Atlas tasks, respectively, which are set to run on 4 and 2 threads, respectively. Thus those first 3 lines actually represent the potential use of, not 12, but 20 threads. If I was actually getting stuff from LHC, there would be no problem, but for some reason I cannot grasp, and which you have not yet tried to address, Boinc seems to think that I don't want/need any tasks from LHC at all, except when one completed task is being reported.
The extreme case with those app_config files would be if LHC went down completely, and no new tasks of any kind were available. Then Einstein and Rosetta are allotted up to 16 tasks on 1 thread each -- leaving 6 unused. Do I care? Not really, so long as Boinc asks both projects to continue to deliver the work. If this were to happen, and it appeared that LHC tasks would be unavailable for some time (say, more than 3 days or so), then a quick change to the other two app_config files would be sufficient to take up the slack -- and then change them back once LHC returned.

If you're feeling game, i'd suggest making a copy of all of your app_config files and put them somewhere safe.
Then remove the max_concurrent, project_max_concurrent from all of them, exit & restart.

Umm... no.
In the Boinc user manual go to Client Configuration, and scroll down to the bottom of the section titled "Project-level configuration." This is the section which discusses the app_config file.
At the bottom, you will find this: If you remove app_config.xml, or one of its entries, you must reset the project in order to restore the proper values. However, it is never really necessary to do this just to change the max number of running tasks for any project. In this instance, and in several other parameters, the value '0' represents "unlimited". It is pretty much equivalent to not having the line in there at all.

So that is what I have done, for everything. I also reset the Resource Share values to default, and will leave those for the foreseeable future. When I get all current and near-future bills sorted out, I will be looking for a GPU capable of running the new Einstein LIGO tasks -- 4 of which look like they'll be able to run on a 16GB card. Talk about crazy -- Einstein as you probably know hands out a constant credit value for each task. For the gamma ray pulsars, it is 693 -- but for the LIGO ones, it's 10,000!! Then I can take another look at the Resource Share values, plus it will free up a few CPU threads for other projects, as I will no longer be running the pulsar stuff (I'm a gravitation theorist, not an astrophysicist -- I don't really care too much about pulsars, other than they make pretty pictures when you point the Webb telescope at them :D )

PS,Boinc seems to be working hard to overcome the massive beating my RAC on LHC took over the past few days -- 21 threads are allotted to LHC as I type.
19) Questions and Answers : Preferences : Rosetta uses only 1 core (Message 109304)
Posted 28 May 2024 by hadron
Post:
Currently running tasks are: Rosetta and Einstein 8 each (controlled by app_config files)
In what way are the being controller by app_config files?
Whatever settings you've got there will impact on work Scheduling, as the Scheduler will have to work around those app_config settings.
LHC;
<app_config>
     <app>
        <name>CMS</name>
        <max_concurrent>2</max_concurrent>
    </app>
    <app_version>
        <app_name>CMS</app_name>
        <avg_ncpus>4</avg_ncpus>
        <plan_class>vbox64_mt_mcore_cms</plan_class>
        <cmdline>--nthreads 4</cmdline>
    </app_version>
   <app>
        <name>Theory</name>
        <max_concurrent>8</max_concurrent>
    </app>
    <app>
        <name>ATLAS</name>
        <max_concurrent>2</max_concurrent>
    </app>
    <app_version>
        <app_name>ATLAS</app_name>
        <avg_ncpus>2</avg_ncpus>
        <plan_class>vbox64_mt_mcore_atlas</plan_class>
        <cmdline>--nthreads 2</cmdline>
    </app_version>
</app_config>
Rosetta:
<app_config>
   <project_max_concurrent>8</project_max_concurrent>
</app_config>
Einstein:
<app_config>
   <app>
       <name>hsgamma_FGRP5</name>
       <max_concurrent>8</max_concurrent>
   </app>
   <app>
       <name>einstein_O3MD1</name>
       <max_concurrent>2</max_concurrent>
   </app>
<app>
    <name>einsteinbinary_BRP4G</name>
       <max_concurrent>4</max_concurrent>
</app>
</app_config>

Only FGRP5 is relevant; O3MD1 (running data gathered by LIGO) was replaced by a GPU-only project, and Boinc just turns up its nose at my GPU (RX560 with 2 GB). BRP4G is a binary pulsar project, which doesn't really interest me very much. For that matter, FGRP5 (crunching gamma ray pulsar data) doesn't really interest me either, but one must do something while waiting to afford another GPU, yes? ;)
I really should clear out all the Einstein tasks, remove those two from the config file, and do a reset, but I've rather too lazy to do that :D



I don't know if task run times are at all relevant
Yep, they are.
In theory- A Task on one Project that runs for a month should result in the same Credit as Tasks on another Project that run for only a couple of minutes each, over that month.
But if the long running Task doesn't make use of trickles
<snip>

Not quite what I had in mind, but it makes a great deal of sense. Make use of trickles? Yeah, we'll see that the day after Bezos and Musk donate all their money to the poor. So Boinc really has no way of knowing what credit that lone Theory task will yield, so it basically has to guess what task loads it should grab. For that matter, the same can pretty much be said of the Atlas tasks too, which until recently were running on one thread, now on two. But even with 2 threads, they still take 4 times as long as one of those Einstein tasks.


I'll leave things as they are for now if you think it best, but I have to say, I am sorely tempted to blow that LHC resource share through the roof, say to something like 5000, to see what happens.
Give it until that exceptionally long running Task is done, and hopefully you'll pick up a more usual runtime Task, then see how it behaves.
Otherwise, no more than double the present LHC resource Share value, otherwise LHC will spike up, and the others will fall down, so you'll bump theirs up (or the LHC down again which would be better), then they'll pick up & LHC will fall, so you bump up it's share again, it'll spike the others will fall, rinse and repeat for months till eventually you get tired of it or things get close to what you want.

Which is why I have somehow managed to resist that temptation.
20) Questions and Answers : Preferences : Rosetta uses only 1 core (Message 109302)
Posted 28 May 2024 by hadron
Post:
I'm using 22 threads out of 24 for Boinc. That seems to still leave me plenty of processing power for the rest of the system.
It appears to be plenty- the difference between CPU time & Run time for your Tasks is around 10 min for the 3hr Tasks, which would indicate a moderately busy system if all cores/threads were available to BOINC.


The current days of work settings have been in place only for a few days; I'd been toying with stupidly large values trying to even things out, but I've given up trying -- now Boinc can do what it will do, and I will try to control that as best i can.
If it had some lagre values, then it will take longer for all of that to clear, and the Scheduler to sort things out.
Give it a few more days with the 0.1 & 0.01 values & things should settle down nicely.

If it still doesn't pick up as much LHC work as you would like, then bump up it's Resource Share, and then give it another few days to see what the effect is.



well, it's 3 days now, and things look like they've settled down. I am not satisfied with the results.
Current settings are days of work 0.1/.0l; resource shares LHC 400, Rosetta 200, Einstein 100.
Currently running tasks are: Rosetta and Einstein 8 each (controlled by app_config files); LHC - 1 Atlas (2 cores) and 1 Theory
Job queue 12 Einstein and 1 Rosetta.
Boinc will only pull in a new task from LHC when it reports a finished task; all other times, it won't request new tasks because it "doesn't need" them.
I don't know if task run times are at all relevant, but in case they are:
Theory can take anywhere from half an hour to several days. The one that's running now has been running for 5-1/2 days, and still has over 4 days to complete. It may be one of those all-too-frequent Theory tasks that will run to just under its drop-dead time, then fail with a computation error. However, it's too early right now to be able to tell. Maybe I'll just assume it's gonna die, and kill it off so I can get a new one.
Atlas tasks take 10 or 11 hours to complete, Rosetta 3 hours, and Einstein 1-1/2 to 2 hours. I have no way of knowing how much "real work" each of those represents.
My RAC on LHC is dropping like a rock.

I'll leave things as they are for now if you think it best, but I have to say, I am sorely tempted to blow that LHC resource share through the roof, say to something like 5000, to see what happens.


Next 20



©2025 University of Washington
https://www.bakerlab.org