Message boards : News : Outage notice
Previous · 1 . . . 3 · 4 · 5 · 6 · 7 · 8 · 9 . . . 13 · Next
Author | Message |
---|---|
Brian Nixon Send message Joined: 12 Apr 20 Posts: 293 Credit: 8,432,366 RAC: 0 |
Looks like today’s was only a small batch, as it’s already run out… |
Mr P Hucker Send message Joined: 12 Aug 06 Posts: 1600 Credit: 11,839,945 RAC: 10,299 |
New tasks started loading at 7:30 UTC today. Looks like we're rolling again. There's no reason to expect a continuous workflow, and why should they have to tell you, when they're busy with the science? Just look in the messages tab of Boinctasks, or the event viewer in Boinc Manager, and see it saying "server has no tasks available". Or check the Rosetta home page. There's a big blue number on their homepage telling you how many million tasks are left. We collectively get through half a million a day. I suggest you join an additional project, you can tell Boinc which you prefer, eg. set Rosetta to weighting 1000 and the other to weighting 1 if you like. Then when you see it getting loads from the other project, you know it's just Rosetta that's offline. 1000 and 1 means it will do 1000 times more Rosetta than the other one. You can even set a project to 0, that means it will only get tasks from there if it has to. If you're just trying to do Coronavirus for example, you could join World Community Grid, and select to do only those type of tasks. They do Cancer, TB, African Rainfall, Immunity aswell if you're interested. |
Sid Celery Send message Joined: 11 Feb 08 Posts: 2125 Credit: 41,245,383 RAC: 9,571 |
I don't know why the server status of this project and most others is hiding information The information is only updated every few hours, in which time 100k tasks could be made available and exhausted. As to why it's only updated every few hours, I have no idea, but it is and has been for years. But essentially, you're right. We're available to grab tasks 247. If there are tasks to grab, we get them. If there aren't, we don't. If people are happy to wait for tasks from only this project, fine. If not, get some from elsewhere. It's all so simple I don't know why I'm bothering to say so. And yet, people still ask. Personally, I increased my runtime to 12hrs from 8hrs, so the few I get last me longer and I don't need so many so other people can get a share too. I'm slipping in a few WCG too, so when tasks come available again, the debt can be repaid to Rosetta with less interruption. Again, it's so simple I don't know why I'm bothering to write that. Running, completing and returning tasks promptly is thousands of times more important to the researchers than it is to any of us. If they need any work done, no-one will be more motivated to provide the tasks than the researchers. That's obvious too. And so, we wait. |
Mr P Hucker Send message Joined: 12 Aug 06 Posts: 1600 Credit: 11,839,945 RAC: 10,299 |
I don't know why the server status of this project and most others is hiding information Update frequency is irrelevant. The problem is the server status page always says 30,000 available, which is wrong. The main homepage can say something like 15,000,000! I can only assume there's a high speed buffer which is what downloads the tasks to us, and that buffer holds 30,000. But it's more sensible to tell us the number of work units in the whole buffer, which is up to about 15,000,000. That number is already publicly available on the homepage, so why not put it in server status too? Personally, I increased my runtime to 12hrs from 8hrs, so the few I get last me longer and I don't need so many so other people can get a share too. I'm not sure what the best to set that to is. They seem to think 8 hours is right so I leave it at that. Fast CPUs will do more work in that 8 hours so there must be returned units with varying sizes of work done. I read something on this forum that they don't need each unit "completed", but each unit helps build a map along with the others. So I'm not sure your extra 4 hours is of any use to them. You're just increasing the resolution in a particular point which they don't need, or they'd be setting the default to 12 hours. I'm slipping in a few WCG too, so when tasks come available again, the debt can be repaid to Rosetta with less interruption. I like WCG because they have a handful of projects and you can pick which you feel is most important. I've got it running cancer research on 62 cores. Running, completing and returning tasks promptly is thousands of times more important to the researchers than it is to any of us. Indeed. There seem to be a lot of people doing this as some kind of competition to see who can get the most points. I do it to further research. It's not worth spending money on electricity just for a number of points, but it is to donate to science. I use the points only as a guide to how well I'm doing and if something isn't working properly. I'll sometimes buy extra hardware because I'm not doing well enough compared to others, but it's for the science not the world position. And sometimes I'll see I'm getting no points on a project and realise something's up - like LHC not returning valid units due to a Virtualbox error. |
sombraguerrero Send message Joined: 5 Mar 18 Posts: 8 Credit: 617,193 RAC: 0 |
I think distributive computation projects have become commonplace enough now that they're starting to take up residence in the collective psyche. Especially right now with the pandemic, it's easy to start viewing something like this as just a part of your daily routine. People don't like having their routines disrupted and so when they are, they want justification for it as they would for anything else. I think that's what people get lost in. It's also easy to get lost in the stats competition as something much simpler and possibly more concrete to some than the underlying science. It is a really good metric for certain aspects of system performance, but I think it's important to have a balance between that and healthy reverence for the larger purpose. I appreciate the stats as a data point. I even preserve them beyond the XMLs so I can record them over broader history, but I mostly like to think of it in terms of, "My computer's on all the time right now anyway. Might as well do the most good I can with the cycles." I for one enjoy the science, but I also think I just appreciate being kept informed enough to know that there isn't a technical difficulty preventing work from coming, because if I know that, I can plan to do other things. |
Mr P Hucker Send message Joined: 12 Aug 06 Posts: 1600 Credit: 11,839,945 RAC: 10,299 |
I think distributive computation projects have become commonplace enough now that they're starting to take up residence in the collective psyche. Especially right now with the pandemic, it's easy to start viewing something like this as just a part of your daily routine. People don't like having their routines disrupted and so when they are, they want justification for it as they would for anything else. I think that's what people get lost in. It's also easy to get lost in the stats competition as something much simpler and possibly more concrete to some than the underlying science. It is a really good metric for certain aspects of system performance, but I think it's important to have a balance between that and healthy reverence for the larger purpose. I appreciate the stats as a data point. I even preserve them beyond the XMLs so I can record them over broader history, but I mostly like to think of it in terms of, "My computer's on all the time right now anyway. Might as well do the most good I can with the cycles." I for one enjoy the science, but I also think I just appreciate being kept informed enough to know that there isn't a technical difficulty preventing work from coming, because if I know that, I can plan to do other things. As long as you do at least 2 projects, you don't get that interruption. You just see it doing different work today and immediately know that project A is offline. If there's no work at all, the fault's at your end. I've joined 8 projects plus 2 beta projects plus 1 "silent" project (as in rarely has work) on 6 computers. I've set the weighting for each project equal to my world position, so the ones I'm behind on get more work, and I know how well I'm doing compared to others. It's an incentive to buy more hardware! But I chose those projects because I think they're interesting and worthwhile, and when I see a task running that interests me, I'll switch the other projects off (no new work setting) and let it do loads of that. At the moment I've got all the CPUs on cancer research, unfortunately GPUs can't do that. |
sombraguerrero Send message Joined: 5 Mar 18 Posts: 8 Credit: 617,193 RAC: 0 |
Oh, I do multiple projects, to be clear, but I still like to know when one is going to be legitimately down so I can adjust some related processes I have. |
Mr P Hucker Send message Joined: 12 Aug 06 Posts: 1600 Credit: 11,839,945 RAC: 10,299 |
Oh, I do multiple projects, to be clear, but I still like to know when one is going to be legitimately down so I can adjust some related processes I have. What is there to adjust? Boinc does that for you. |
sombraguerrero Send message Joined: 5 Mar 18 Posts: 8 Credit: 617,193 RAC: 0 |
I have an ETL I run on my stats data to do my own stuff with it, for fun, and for education. |
Mr P Hucker Send message Joined: 12 Aug 06 Posts: 1600 Credit: 11,839,945 RAC: 10,299 |
I have an ETL I run on my stats data to do my own stuff with it, for fun, and for education. Then upgrade it so it can handle the common thing that is server outages and work shortage. |
sombraguerrero Send message Joined: 5 Mar 18 Posts: 8 Credit: 617,193 RAC: 0 |
Oh I have, but because outages really just leave the data stale, quite often, I'd rather modify the consumption than stop it altogether, especially when it comes to spanning multiple projects. The point I'm trying to make with this is that there are legitimate reasons for wanting to know if a project is actually in a fail state. Most of the time, they aren't, but it's nice to know when they are. |
Jim1348 Send message Joined: 19 Jan 06 Posts: 881 Credit: 52,257,545 RAC: 0 |
The point I'm trying to make with this is that there are legitimate reasons for wanting to know if a project is actually in a fail state. Of course. Most projects make some attempt to notify you when there is a planned maintenance outage or a lack of work. Rosetta doesn't. Several types of work units have some conflict with each other, so that running them together slows down the total output. I can't plan on how to juggle that with Rosetta. I had four machines (two Ryzen 3900X and two Ryzen 3950X) on Rosetta until recently, Now I am down to zero. I am actually glad that they can get their work done without so many crunchers, if that is what they want. If it is not what they want, they need to wake up and say something. I don't need to explain to them. They need to explain to me. |
Sid Celery Send message Joined: 11 Feb 08 Posts: 2125 Credit: 41,245,383 RAC: 9,571 |
I don't know why the server status of this project and most others is hiding information I don't know if it's a high-speed buffer. It has been explained but I didn't care enough to be paying attention tbh. I thought it was something to do with being converted into a downloadable format for us, but I'm probably wrong. Personally, I increased my runtime to 12hrs from 8hrs, so the few I get last me longer and I don't need so many so other people can get a share too. It returns more decoys and I think that makes a difference. I'm not sure what the best setting is either - I'm just eking out the tasks and returning more data while there's a current shortage, which now seems to be solved. I'll keep it at 12hrs until Sunday, just to make sure the current supply isn't only temporary, then I'll switch back to the default 8hrs. I'm slipping in a few WCG too, so when tasks come available again, the debt can be repaid to Rosetta with less interruption. I do too. I haven't limited my tasks only to Open Pandemics. I don't see any of their projects as any less important - probably the other parts of the project need some extra attention while everyone else is focussing on CV19 for the more obvious reasons. Indeed. There seem to be a lot of people doing this as some kind of competition to see who can get the most points. I do it to further research. It's not worth spending money on electricity just for a number of points, but it is to donate to science. I use the points only as a guide to how well I'm doing and if something isn't working properly. I'll sometimes buy extra hardware because I'm not doing well enough compared to others, but it's for the science, not the world position. And sometimes I'll see I'm getting no points on a project and realise something's up - like LHC not returning valid units due to a Virtualbox error. Nicely said |
Mr P Hucker Send message Joined: 12 Aug 06 Posts: 1600 Credit: 11,839,945 RAC: 10,299 |
The point I'm trying to make with this is that there are legitimate reasons for wanting to know if a project is actually in a fail state. Oh for goodness sake. If you can't handle a varying workload you're doing something drastically wrong. I manage just fine. Find a backup project. |
Mr P Hucker Send message Joined: 12 Aug 06 Posts: 1600 Credit: 11,839,945 RAC: 10,299 |
Update frequency is irrelevant. The problem is the server status page always says 30,000 available, which is wrong. The main homepage can say something like 15,000,000! I can only assume there's a high speed buffer which is what downloads the tasks to us, and that buffer holds 30,000. But it's more sensible to tell us the number of work units in the whole buffer, which is up to about 15,000,000. That number is already publicly available on the homepage, so why not put it in server status too? Perhaps. But I think all projects should show the larger figure (or both) on the server status page. what the users want to know is how much work is left to come. Most projects don't even list the larger figure anywhere, at least Rosetta puts it on the homepage. I'm not sure what the best is to set it to either. They seem to think 8 hours is right so I leave it at that. Fast CPUs will do more work in that 8 hours so there must be returned units with varying sizes of work done. I read something on this forum that they don't need each unit "completed", but each unit helps build a map along with the others. So I'm not sure your extra 4 hours is of any use to them. You're just increasing the resolution in a particular point which they don't need, or they'd be setting the default to 12 hours. If they don't need those extra decoys, you're wasting CPU time. Probably better using it for another similar project. If they do need the extra decoys, they would have set the 8 hours higher themselves. I like WCG because they have a handful of projects and you can pick which you feel is most important. I've got it running cancer research on 62 cores. It probably evens out actually. Anyone (probably quite a lot of people) with their settings on the default of "take anything that's there" will get whatever others don't select. So people saying "I'm only doing Coronavirus (I refuse to use the technical name) work" aren't actually making it happen faster, because they "do anything" users just end up getting less of it as they've taken WUs out of the queue. |
sombraguerrero Send message Joined: 5 Mar 18 Posts: 8 Credit: 617,193 RAC: 0 |
Several types of work units have some conflict with each other, so that running them together slows down the total output. I can't plan on how to juggle that with Rosetta. To this point, I feel like Rosetta is more than a bit hoggish at times of the resource share, in spite of the client purportedly distributing the work load evenly. All other project WUs seem quite outnumbered and overpowered when Rosetta is active on my box. |
Grant (SSSF) Send message Joined: 28 Mar 20 Posts: 1683 Credit: 17,919,030 RAC: 22,447 |
To this point, I feel like Rosetta is more than a bit hoggish at times of the resource share, in spite of the client purportedly distributing the work load evenly.It doesn't distribute it evenly, it distributes it in accordance with your Resource share settings. Your Average Credit for your projects shows that Rosetta is getting the least amount of processing time, so how your other projects can be "outnumbered and overpowered when Rosetta is active" makes no sense to me at all. Since you are using your GPU, you should reserve a CPU core to support each running GPU WU in order to avoid any impact on GPU work with the CPU trying to use all CPU cores and threads for CPU work as well as supporting the GPU. Grant Darwin NT |
sombraguerrero Send message Joined: 5 Mar 18 Posts: 8 Credit: 617,193 RAC: 0 |
It does distribute it evenly if you never touch anything, hehe. I only ever do have one gpu wu running at one time, so that's not too much of an issue. I think it's just surprised me over time how little Rosetta yields for the default duration it's expected to cycle. I also find the way the Boinc projects handle the resource share settings a little too arbitrary, but everyone has their own sweet spot, I suppose. In my case, even with the hour allocation set lower, I always get assigned really long running Rosetta units and I honestly would rather spend that time on other projects anyway, now that I've been tracking stats for a good while. |
Grant (SSSF) Send message Joined: 28 Mar 20 Posts: 1683 Credit: 17,919,030 RAC: 22,447 |
I think it's just surprised me over time how little Rosetta yields for the default duration it's expected to cycle.Or that the other projects you crunch for pay more than they should. If all projects awarded Credit according to the definition of the Cobblestone, then the differences between projects would be minimal. But they don't- so some pay the going rate, others less, some more & and there are those that pay stupidly obscene amounts more. I always get assigned really long running Rosetta unitsAs does everyone. It is not like other projects where the time taken to process a task varies depending on the capability of the system, the default processing time for Rosetta Tasks is 8 hours. Some may bail out early, others may take up to 10 hours longer. But the vast majority run to the set Target CPU time. Grant Darwin NT |
sombraguerrero Send message Joined: 5 Mar 18 Posts: 8 Credit: 617,193 RAC: 0 |
That makes sense in the context of the stats then. I guess i was just making assumptions based on past experience. I thank you for the input, sir! |
Message boards :
News :
Outage notice
©2024 University of Washington
https://www.bakerlab.org