Message boards : Number crunching : Tells us your thoughts on granting credit for large protein, long-running tasks
Previous · 1 · 2 · 3 · 4 · 5 . . . 9 · Next
Author | Message |
---|---|
Sid Celery Send message Joined: 11 Feb 08 Posts: 2119 Credit: 41,179,074 RAC: 11,480 |
The answer is +25% credit compensation. Interesting. To throw a few more left-field ideas out there... 1) No change to whatever credits are normally awarded, but award some variation of additional new "Bad Mutha" badges 2) Set a cookie that allows a discount at approved hardware suppliers on purchase of CPU/RAM with some kind of commission also going back to Rosetta/Bakerlab/IPD I know this isn't a commercial project but there must be some way of monetising this stuff for the benefit of both researchers and volunteers ...don't shoot the messenger |
Tomcat雄猫 Send message Joined: 20 Dec 14 Posts: 180 Credit: 5,386,173 RAC: 0 |
Personally, I think the best option would be to try and balance these tasks to grant about as many credits per hour as "normal" tasks. Since, for me, Rosetta Mini has consistently returned 2-6X the credits of regular Rosetta tasks and has been way more consistent (don't know about others), maybe balance the large tasks to match the payout of Rosetta Mini. I don't think there should be a bonus unless one can directly control whether these tasks come in or not (as in, having a toggle in the Rosetta preference panel for big bad proteins). If such a toggle was added, methinks a 10% bonus is reasonable. Methinks it would be slightly more beneficial if one was granted a bonus for having longer run times. Like a 2% bonus for tasks which ran above the default runtime and a 5% bonus for tasks ran for over a day. Since the watchdog has been extended, maybe grant an additional 0-10% compensative bonus for tasks that ran significantly over the set runtime? |
teacup_DPC Send message Joined: 3 Apr 20 Posts: 6 Credit: 2,744,282 RAC: 0 |
Unread Message 94951 - Posted: 19 Apr 2020, 23:16:02 UTC - in response to Message 94937. I think your remark is valid when we leave from the assumption that systems able to do 4GB tasks without any issue besides doing other things are rare. To be honest when writing my reply I was not sure about that. That is why I wrote this sentence: Where exactly to position the credits between 1x1GB and 4x1GB depends on the availability of processor cores and memory in the clients capable for the 4GB jobs, you can judge that better than I.Another thing I am not sure about is that 4GB WU's will take more time to solve. It seems logical, while more data needs to be moved around. This would imply there is not only a lower limit on memory size, but on core performance as well. If so, where does the threshold lie? From one side I expect however you will have a point. Lets look at an extremity of the spectrum. Ryzen 9 3950X, 16core/32thread. When you only want to harvest the 4GB jobs for each thread, more than 4x32=128GB of memory is needed. Systems like this will be a minority, I expect. But besides this extremity what about an average desktop, 8 years or younger, quite some systems will have 16GB, a majority will have 8GB of RAM, a minority will have 32GB or more. All those systems are able to do without much effort, besides doing other things, at least 1 4GB WU. Assuming their processor cores can cope with the load. So back to the question, how probable are those systems? Perhaps Mod.Sense can shed some light on this? |
Ged Send message Joined: 17 Apr 06 Posts: 2 Credit: 1,034,115 RAC: 0 |
For me, personally, I'm not driven by the credits granted for running work units; It's about contributing to the science, either by running work units which model a particular behaviour or sheer crunching of data for further treatment or research candidate selection/rejection. I'd rather the see application development *and* testing effort be expended producing efficient and effective code. I'd also like to see more realistic operational criteria being assigned to work units so as not to 'waste' computing effort (and electricity) by having my machines swamped with, often, spuriously defined deadlines, maybe by including some operational acceptance testing rather than just functional tests. That's my 10c's worth ;-) Ged |
Tomcat雄猫 Send message Joined: 20 Dec 14 Posts: 180 Credit: 5,386,173 RAC: 0 |
Personally, I think the best option would be to try and balance these tasks to grant about as many credits per hour as "normal" tasks. Since, for me, Rosetta Mini has consistently returned 2-6X the credits of regular Rosetta tasks and has been way more consistent (don't know about others), maybe balance the large tasks to match the payout of Rosetta Mini. I don't think there should be a bonus unless one can directly control whether these tasks come in or not (as in, having a toggle in the Rosetta preference panel for big bad proteins). If such a toggle was added, methinks a 10% bonus is reasonable. Welp, my problem with Rosetta mini generating way more credits than regular Rosetta tasks has been fixed (apparently my computer was generating more than the max allowed credit? Oops) I still hold the belief that no bonus is necessary for this task, unless there is a setting in the Rosetta preference panel that specifically toggles "big bad proteins". Even if such tasks carried quite a large bonus, I doubt anyone would upgrade their computer just for these tasks. |
Grant (SSSF) Send message Joined: 28 Mar 20 Posts: 1675 Credit: 17,738,985 RAC: 22,900 |
I still hold the belief that no bonus is necessary for this task, unless there is a setting in the Rosetta preference panel that specifically toggles "big bad proteins". Even if such tasks carried quite a large bonus, I doubt anyone would upgrade their computer just for these tasks.The proposal for more Credit isn't to get people to upgrade their computers, it's just so that the people that do process them don't lose out on Credit because they will be unable to process as many as 3 other Tasks at the same time as one of the very large RAM requirement Tasks is running (actually under certain circumstances some people may not be able to process as many as 5 other Tasks till the vary large RAM requirement Task is done, or at the very least during various stages of it's progress). Grant Darwin NT |
Mod.Sense Volunteer moderator Send message Joined: 22 Aug 06 Posts: 4018 Credit: 0 RAC: 0 |
For me, personally, I'm not driven by the credits granted for running work units; It's about contributing to the science, either by running work units which model a particular behaviour or sheer crunching of data for further treatment or research candidate selection/rejection. Ged, I just wanted to clarify, are you basically suggesting that you'd like to see some way to control the deadline of the work you receive? Or have a way to only be assigned WUs that have 8 day deadlines? Or are you referring to cases where the BOINC Manager gets tricked into requesting more R@h work than is required to fill your work cache, and to complete before the 3 day deadlines? Rosetta Moderator: Mod.Sense |
Tomcat雄猫 Send message Joined: 20 Dec 14 Posts: 180 Credit: 5,386,173 RAC: 0 |
I still hold the belief that no bonus is necessary for this task, unless there is a setting in the Rosetta preference panel that specifically toggles "big bad proteins". Even if such tasks carried quite a large bonus, I doubt anyone would upgrade their computer just for these tasks.The proposal for more Credit isn't to get people to upgrade their computers, it's just so that the people that do process them don't lose out on Credit because they will be unable to process as many as 3 other Tasks at the same time as one of the very large RAM requirement Tasks is running (actually under certain circumstances some people may not be able to process as many as 5 other Tasks till the vary large RAM requirement Task is done, or at the very least during various stages of it's progress). Ah, that's a really good point. Didn't think of it that way. If that is the case, I think it might be fair to give quite a large bonus for these tasks. Since we don't know how much more RAM these tasks take on average, it's still hard to tell exactly how big a bonus they should be given. Unless there is a way to take the RAM usage into account when calculating credits. For example It might be fair to give a big bad protein task that ate up twice the ram that a regular Rosetta task does a 100% credit bonus. We need someone with a economics background to weigh in, this seems like something that involves opportunity costs and whatnot. |
Grant (SSSF) Send message Joined: 28 Mar 20 Posts: 1675 Credit: 17,738,985 RAC: 22,900 |
We need someone with a economics background to weigh in*shudder* Things are messy enough as it is. Economists are like lawyers, ask 20 different ones something and you'll get 20 different answers (or the the answer you want if you pay for it). *another shudder* Grant Darwin NT |
Tomcat雄猫 Send message Joined: 20 Dec 14 Posts: 180 Credit: 5,386,173 RAC: 0 |
We need someone with a economics background to weigh in*shudder* *sigh* I guess there is a reason economics is called the "dismal science". I still think the goal is to balance these tasks so running them don't affect one's RAC by much. Since these tasks can take up to 4GBs instead of the regular 1GB (is it?), maybe a 150% bonus is fair? |
Grant (SSSF) Send message Joined: 28 Mar 20 Posts: 1675 Credit: 17,738,985 RAC: 22,900 |
*sigh* I guess there is a reason economics is called the "dismal science". I still think the goal is to balance these tasks so running them don't affect one's RAC by much. Since these tasks can take up to 4GBs instead of the regular 1GB (is it?), maybe a 150% bonus is fair?Since it's reducing their output by 75%, triple the going rate would be fair (eg 4 Tasks, 100 each, lose 3 that's 300 lost. One Task *3 =300, still less than the amount lost, but not nearly as much). Of course this would give a boost in Credit for 1 or 2 core system with lots of RAM, or those huge multi core/thread systems with extreme amounts of RAM. If more people are affected negatively by the new Tasks than benefit (which i suspect will be the case), then the higher level of Credit should be paid. Otherwise still give a bonus for those Tasks, but not as large a one. And the impact will depend on just how much of the total processing time the Tasks require that much RAM. If it's for 5% or 10% of a run, then the impact of lost output won't be significant at all. But if it's for 25% or more of the time, then that's a big hit for those affected. Grant Darwin NT |
Mod.Sense Volunteer moderator Send message Joined: 22 Aug 06 Posts: 4018 Credit: 0 RAC: 0 |
Admin posted that the average model in their lab actually used under 2GB. The 4GB is the maximum the WU it ALLOWED to use. It is like asking how much water a Tesla gigafactory is going to use, and looking at the size of the pipe that goes in to the factory. But the system must be sized large enough to run the sprinkler system in case of fire. It doesn't mean the maximum flow is used all of the time. Rosetta Moderator: Mod.Sense |
Tomcat雄猫 Send message Joined: 20 Dec 14 Posts: 180 Credit: 5,386,173 RAC: 0 |
Admin posted that the average model in their lab actually used under 2GB. The 4GB is the maximum the WU it ALLOWED to use. Hmm, how much do regular Rosetta tasks use on average? |
Grant (SSSF) Send message Joined: 28 Mar 20 Posts: 1675 Credit: 17,738,985 RAC: 22,900 |
Hmm, how much do regular Rosetta tasks use on average?I've seen from 80MB to 1.5GB. Often around 400MB-800MB. And i recently found out about wuprop.boinc-af.org. You need to checkout the graphs for the full story, and it looks like there already plenty of Tasks already hitting (or at least requesting) 2GB of RAM (don't know what the source is- used or requested RAM?). Grant Darwin NT |
Mod.Sense Volunteer moderator Send message Joined: 22 Aug 06 Posts: 4018 Credit: 0 RAC: 0 |
Grant's link didn't work for me, try this one Very cool! A BOINC project to collect & process project stats. Rosetta Moderator: Mod.Sense |
Tomcat雄猫 Send message Joined: 20 Dec 14 Posts: 180 Credit: 5,386,173 RAC: 0 |
Hmm, how much do regular Rosetta tasks use on average?I've seen from 80MB to 1.5GB. Often around 400MB-800MB. Thanks, if that is the case, I'll assume big bad proteins would require close to 2X the RAM of a regular protein on average, then. Assuming that these tasks generate a similar amount of credits per core per hour, it would seem fair to give theses big bad proteins a 100% bonus. However, that is a worst-case scenario estimate, made under the assumption that the user is running at maximum RAM usage and each big bad protein will take up two slots. Some users may be able to run as many BBPs as regular proteins (best case scenario, no bonus needed). I'll assume that from worst to best case, users on Rosetta follow a normal distribution. So I guess a 50% bonus should be a good starting point. |
bkil Send message Joined: 11 Jan 20 Posts: 97 Credit: 4,433,288 RAC: 0 |
Folding@home offered various bonuses during its lifetime, I think they had beta bonus for completing WUs that were not correctly calibrated yet or may crash, big bonus for upper-end requirement outliers, bigadv for those tasks requiring lots of cores, lots of runtime and lots of RAM, and a quick return bonus if for completing short deadlines and running 24/7.
- https://foldingathome.org/support/faq/points/
|
Sid Celery Send message Joined: 11 Feb 08 Posts: 2119 Credit: 41,179,074 RAC: 11,480 |
We need someone with a economics background to weigh in*shudder* *tut* I already solved this: 1) No change to whatever credits are normally awarded, but award some variation of additional new "Bad Mutha" badges 1 is for people who don't care about credits 2 is for those who want 'paying' in some sense And my earlier one is proving more apt the more messages I read. Someone always wants to create a pretence of 'rationality' about credits, while suggesting something that's little more than a different version of 'random' Pick a number. It's (barely) funny coz it's true |
Sid Celery Send message Joined: 11 Feb 08 Posts: 2119 Credit: 41,179,074 RAC: 11,480 |
Hmm, how much do regular Rosetta tasks use on average?I've seen from 80MB to 1.5GB. Often around 400MB-800MB. When we were getting all those 1.5Gb tasks and RAM was pushing against my 16Gb RAM I spotted one of 2.333Gb, Just the once though and a couple of weeks ago now |
Grant (SSSF) Send message Joined: 28 Mar 20 Posts: 1675 Credit: 17,738,985 RAC: 22,900 |
We only have a few top notch 32+ cores machines with beefy GPUs around the world,Try hundreds of thousands, at the least. but if we contributed every phone, tablet and low-mid end office machine, typically with 2-4 cores, our computing capacity could increase by orders of magnitude. (I.e., we have way less than a million hosts and there exist billions of personal computing devices in the world)For as many of of those devices there are, many are of such low capability they are of no use to many projects. And for those that are of use, their frequent use for what they were designed for by the users means they often can't contribute much during those periods, compared to more capable systems. And you need to keep in mind efficiency isn't actually about low peak or maximum power use- it is about energy used over time to complete a task. It's no good having a device use 1W if it takes 1 month to produce a result when something that uses 1kW can produce the same result in a matter of seconds. Yeah, it's instantaneous power consumption is a lot higher. But it uses less energy to do the same work. And the fact it can do so much more work over the same period of time as the slower device makes it even more useful to a project. Grant Darwin NT |
Message boards :
Number crunching :
Tells us your thoughts on granting credit for large protein, long-running tasks
©2024 University of Washington
https://www.bakerlab.org