Index | Recent Threads | Unanswered Threads | Who's Active | Guidelines | Search |
![]() |
World Community Grid Forums
![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() |
No member browsing this thread |
Thread Status: Active Total posts in this thread: 100
|
![]() |
Author |
|
SekeRob
Master Cruncher Joined: Jan 7, 2013 Post Count: 2741 Status: Offline |
Since I stepped from 4.4 to 4.13 with every major release, including 4.10 and the screen problem remained, but the HWE command did solve the issue, it's something outwith the kernel i.e. the sampled command on the HW front that seems to have fixed things. At any rate... humming along in seeming stable performance, no tailing off.
Some googling suggests there were issues with the early 16.04 v.v. I7 6700 |
||
|
KerSamson
Master Cruncher Switzerland Joined: Jan 29, 2007 Post Count: 1673 Status: Offline Project Badges: ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() |
Hi SekeRob,
----------------------------------------I've just updated to 4.10.33 incl. xserver. Everything is running well ... excepted that the login screen does not appear, only the login background picture is showed. Currently, I have only a remote CLI connection over SSH. But the system is crunching. Yves ---------------------------------------- [Edit 1 times, last edit by KerSamson at Sep 12, 2017 8:25:26 PM] |
||
|
Sgt.Joe
Ace Cruncher USA Joined: Jul 4, 2006 Post Count: 7668 Status: Offline Project Badges: ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() |
I have an observation which is not scientifically rigorous, but appears on the surface to be interesting. I have a 2600K system with 8 gb of memory running Linux which had running both SCC and MIP. The average time gradually rose over the days and appeared to settle about 3.5 to 4 hours per MIP unit. I switched it over to all MIP to see if I had any of the symptoms described by KerSamson. Once it was doing only MIP work the average work unit was taking 5.5 to 6.5 hours. I thought maybe it was the result of just bigger workunits. I also have a system running all MIP units, a dual Xeon E5410 with 8gb of memory also running Linux. This machine should be slower than the 2600K, but even when the 2600K was running the longer units, the Xeon continued to run units in the 4 to 4.5 hour range with only an occasional unit being longer.
----------------------------------------I don't know what exactly is happening, but suspect the Xeon is a bit better at running the MIP units than the 2600K. There must be some piece of the architecture in the 2600k which is being overloaded versus the older Xeon. The Xeon may have some part of its architecture which is less prone to being maxed out than the 2600K. It may have something to do with the 2600K being a consumer chip and the Xeon being a server chip. When both machines were running SCC exclusively the 2600K was notably faster by about 15 to 20 minutes per work unit. This is the only time I have noticed this kind of difference in the performance of the two machines. I have since switched the 2600K back to SCC. Just an observation. Cheers
Sgt. Joe
*Minnesota Crunchers* |
||
|
TonyEllis
Senior Cruncher Australia Joined: Jul 9, 2008 Post Count: 261 Status: Offline Project Badges: ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() |
Interesting, wonder if this difference has anything to do with both working memory size and effectiveness of Hyper-threaded Cores for the MIP workload...
----------------------------------------2600k - 8 CPU (4 cores hyper-threading) 8M L2 cache - assuming Hyper-threading turned on dual Xeon E5410 - 8 CPU (4+4 - no Hyper-threading available) 24M L2 cache (12+12) Hyper-threading does not give anywhere near a 100% increase (with a few workloads it can actually decrease productivity) - so would expect the 8 CPU combined of the Xeons to be more efficient than 4 cores Hyper-threaded to provide 8 CPU + a big difference in L2 cache... My Ivy Bridge i7-3770 - CPU time while running MIP exclusively varied between 1.92 and 3.02 hours - didn't notice a slowdown... Linux kernel 2.6.32-696.v6.x86_64 Ivy Bridge i3-3220 - CPU time running MIP exclusively varies between 1.66 and 3.64 hours - no slowdown evident... Linux kernel 3.9.10-100.fc17.x86_64 both have 8G memory i7-3770 - 8 CPU L2 8M cache clocked at 2.9 - 3.2 Mhz i3-3220 - 4 CPU L2 3M cache clocked at 3.3 MHz Not a substantial difference in L2/CPU - the i7 Intel supplied heatsink/fan is inadequate and will be replaced by a Hyper 212 II EVO - hence the lower clock speeds to prevent overheating... Workload other than WCG on the i3 varies much more than that on the i7.
Run Time Stats https://grassmere-productions.no-ip.biz/
|
||
|
Sgt.Joe
Ace Cruncher USA Joined: Jul 4, 2006 Post Count: 7668 Status: Offline Project Badges: ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() |
Yes, I have hyperthreading turned on in the 2600K. You could be right on either the hyperthreading issue or the L2 cache difference or may a combination of both items. MIP does something different than SCC and so far MIP is the only project where the Xeon is faster than the 2600K. Thanks for the input.
----------------------------------------Cheers
Sgt. Joe
*Minnesota Crunchers* |
||
|
TonyEllis
Senior Cruncher Australia Joined: Jul 9, 2008 Post Count: 261 Status: Offline Project Badges: ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() |
Managed to obtain a mip1 running on a machine running mostly scc1 - so checked memory usage (Linux)...
----------------------------------------[root@danda ~]# ps -o rss,sz,vsz 5017 1639 11174 RSS SZ VSZ 309240 122705 490820 # mip1 51036 31971 127884 # scc1 55856 33183 132732 # scc1 [root@sophia ~]# ps -o rss,sz,vsz 4779 RSS SZ VSZ 301644 98652 394608 # mip1 RSS: resident set size, the non-swapped physical memory that a task has used (in kiloBytes). SZ: size in physical pages of the core image of the process. This includes text, data, and stack space. VSZ: virtual memory size of the process in KiB (1024-byte units). Substantial difference between mip1 and scc1 ( added comments to output to indicate project)
Run Time Stats https://grassmere-productions.no-ip.biz/
----------------------------------------[Edit 1 times, last edit by TonyEllis at Sep 14, 2017 2:47:59 AM] |
||
|
der_Day
Cruncher Joined: Feb 14, 2008 Post Count: 2 Status: Offline Project Badges: ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() |
can anyone say something about the checkpoints? 1 checkpoint after <10minutes and the next after another hour?! I crunch with a private computer, so I have to suspend it or are there other tricks?
----------------------------------------[Edit 1 times, last edit by der_Day at Sep 14, 2017 5:29:48 PM] |
||
|
smeyer55
Senior Cruncher Joined: Feb 15, 2009 Post Count: 303 Status: Offline Project Badges: ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() |
Another observation on a Windows 10 machine running a ThreadRipper. It was running MIP exclusively and I saw the gradual rise in run times over a few days to 5-6 hours each. CPU temp also fell as KerSamson reported. CPU usage showed as 100% during this time.
After reading these posts I set to run SCC and MIP together. The MIP run times have fallen back to the 2.2 hour times. So there is definitely something odd happening when running only MIP on a hyperthreading machine. |
||
|
Sgt.Joe
Ace Cruncher USA Joined: Jul 4, 2006 Post Count: 7668 Status: Offline Project Badges: ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() ![]() |
can anyone say something about the checkpoints? 1 checkpoint after <10minutes and the next after another hour?! I crunch with a private computer, so I have to suspend it or are there other tricks? I think the checkpoints are hard coded into the application, so there is probably nothing you can do to change when they occur.(Someone correct me if I am wrong.) You can check all the projects and run just ones that have checkpoints which occur at regular short intervals. Cheers
Sgt. Joe
*Minnesota Crunchers* |
||
|
Former Member
Cruncher Joined: May 22, 2018 Post Count: 0 Status: Offline |
I have to support what smeyer55 has posted except I'm not monitoring CPU temps. Anecdotally (Just visual observation) it looks like the greater the number of threads the slower the WUs run. My 4 thread machines run them faster than the 8 thread, which run them faster than the 16 thread etc. It takes almost 7 hours for one to complete on a 32 thread machine which has faster processors than the 16 thread. after I get to a certain badge level, I may use app_config to limit MIP1 to 25% of the threads in a machine. It would be interesting to profile this app with valgrind looking at cache misses or branch prediction misses etc.
|
||
|
|
![]() |