Index | Recent Threads | Unanswered Threads | Who's Active | Guidelines | Search |
![]() |
World Community Grid Forums
![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() |
No member browsing this thread |
Thread Status: Active Total posts in this thread: 57
|
![]() |
Author |
|
Former Member
Cruncher Joined: May 22, 2018 Post Count: 0 Status: Offline |
Afterthought for ingleside: Whilst the proposed validator load splitting strikes on the face as logical and efficient, the results for all sciences seem to run one long series... we're now somewhere at 616 million originals, before splitting them for quorum [You see those numbers running past when updating with WCGDAWS]. Just wondered how that works when there are multiple sciences, a validator or validators per science. Feel like these could be running cycles to find out, so guess there's working off some subset table or secondary indices to find which ones they should be looking at.
|
||
|
knreed
Former World Community Grid Tech Joined: Nov 8, 2004 Post Count: 4504 Status: Offline Project Badges: ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() |
Here are the backend daemons that are currently running:
Server #1 Variables: --d sets the level of logging information --sleep_interval setsthe number of seconds to wait before querying the database again for those rare times where the previous query returned nothing to do --mod X Y means process workunit.id % X == Y (or result.id % X == Y for those daemons operating on the result table) --min_age_days sets the number of days before deleting a workunit after all of its files have been deleted |
||
|
Former Member
Cruncher Joined: May 22, 2018 Post Count: 0 Status: Offline |
An interesting number and distribution, 8 hcc1 assimilators spread over 2 servers, and 2 assimilators for sn2s, where all others but hcc1 have 1 (odd in result quantitative terms compared to other sciences).
(Show the back of your tongue and we see what you're eating too ;) Thanks |
||
|
Former Member
Cruncher Joined: May 22, 2018 Post Count: 0 Status: Offline |
And this from the Known Issues DDP thread:
We are caught up now and work is flowing freely. In order to help volunteers keep their machines contributing during these outages, we have expanded some setting that control how much can be cached. We are now using the following settings" The daily quota time multiplier is what a device is getting at most for a given resource, so 2 GPU's would be 300*15*2 = 9,000 a day. If 1 is more powerful than the other, the processing distribution could be different... the servers do not care FAIK. 2 cards of unequal make and you still got a chance to buffer 2,400... a good few hours. :D<daily_result_quota>300</daily_result_quota>[Jan 16, 2013 7:13:01 PM] |
||
|
themoonscrescent
Veteran Cruncher UK Joined: Jul 1, 2006 Post Count: 1320 Status: Offline Project Badges: ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() |
I'm now in to 43 pages of pending for GPU
----------------------------------------![]() What file does the new control go in? ![]() ![]() |
||
|
Ingleside
Veteran Cruncher Norway Joined: Nov 19, 2005 Post Count: 974 Status: Offline Project Badges: ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() |
Afterthought for ingleside: Whilst the proposed validator load splitting strikes on the face as logical and efficient, the results for all sciences seem to run one long series... we're now somewhere at 616 million originals, before splitting them for quorum [You see those numbers running past when updating with WCGDAWS]. Just wondered how that works when there are multiple sciences, a validator or validators per science. Feel like these could be running cycles to find out, so guess there's working off some subset table or secondary indices to find which ones they should be looking at. Have no experience with databases, but making a guess where'll possibly one index for any wu with the NEED_VALIDATE-flag set, or possibly one index for each application/NEED_VALIDATE-combination. If the former, example FAAH would need to search-through the index until finds a wu with FAAH as the application, while HCC would need to check wuid 1st. and only afterwards check if the application is HCC. If the latter, HCC would need to check the wuid while FAAH would only need to check if anything is present in it's index or otherwise sleep. ![]() "I make so many mistakes. But then just think of all the mistakes I don't make, although I might." |
||
|
Hypernova
Master Cruncher Audaces Fortuna Juvat ! Vaud - Switzerland Joined: Dec 16, 2008 Post Count: 1908 Status: Offline Project Badges: ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() |
Its long ago that I stopped using the "Results" page as it will not display anymore. I estimate that there are probably days were I have over 30'000 results in various states, that means 2000 pages at 15 results per page, or so. This cumulation is also due to validators issues. It seems that it becomes not manageable anymore.
----------------------------------------When I look at the results per device it is becoming very erratic. I know that there were issues with the databases, I hope this will all settle. The GPU crunching has been an excellent stress test of the WCG infrastructure and shows the limitations. Why not put 10 or 20 diffraction images into one WU for HCC. The crunching time will rise and there will be much less WUs to manage. The network bandwith will also be reduced as there will be less frantic traffic. HCC returns per day a number of WUs equivalent to over the sum of all other active projects. With GPU crunching the traffic has doubled. On GPU Grid they are two sizes of WU which are sent according to the graphic card class (all NVidia type here). The powerful GPU cards like the 580GTX or 680GTX receive WUs that may have up to 8 hour runtime each. ![]() [Edit 1 times, last edit by Hypernova at Jan 17, 2013 4:08:38 PM] |
||
|
Hypernova
Master Cruncher Audaces Fortuna Juvat ! Vaud - Switzerland Joined: Dec 16, 2008 Post Count: 1908 Status: Offline Project Badges: ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() |
The last post of knreed regarding the major database issues mentions as a possibility the repacking of WUs. It goes in the right direction. But knreed please if you repacking GPU WUs just tell us well in advance so that we can adapt the app_info.xml file to change the WU type number and avoid having idle times on the dedicated machines.
----------------------------------------![]() |
||
|
Former Member
Cruncher Joined: May 22, 2018 Post Count: 0 Status: Offline |
Re outage announcement: http://www.worldcommunitygrid.org/forums/wcg/viewpostinthread?post=408639 Maybe the "various" word used in noted knreed post on the Jan.22 24 hour long outage [starting 03:00 UTC], was meant to be "varying" sizes. The idea is still to create WU's for sciences [where it's possible!], that are matched to the groups of hosts with different power [Limiting the sample to just for CPU, a centrino duo getting a WU with 2 HCC images, a I7-2600 getting WU's with 10 images or the former getting a FAAH job with 20 dockings and the latter with 100 dockings]. In all, when the average target run time of a science is 6 hours, it will be much closer to that target of 6 hours for all and not the sum of the slowest to the fastest and within runtimes ranging from 2 to 24 hours.
Here's hoping |
||
|
Former Member
Cruncher Joined: May 22, 2018 Post Count: 0 Status: Offline |
BTW, v.v. the 2 outage notices, there's an unsaid piece of good news [unless I missed it being said] in the not so good news. Hope knreed is able to combine this on the go whilst doing the software side of the upgrades. Knowing WCG/IBM policy, let's not get ahead of ourselves, and don't set an expectation that could fly in the face [Mr. Murphy is 24/7 attentive ;O].
|
||
|
|
![]() |