Message boards : Number crunching : 1 Work Unit Per GPU
Author | Message |
---|---|
I would like to suggest a 1 WU per GPU policy for this project in order to increase its efficient use of resources and speed up results. | |
ID: 46314 | Rating: 0 | rate: / Reply Quote | |
For people that are like TL:DR, Set in BOINC 0 days works and 0 additional days work. If you crunch 2 WU at a time per GPU, it's probably not the best time to do that, especially because the WU have such high utilization now. We just want to have more people crunching the WUs in parallel. | |
ID: 46316 | Rating: 0 | rate: / Reply Quote | |
I also agree that the scientists should implement, at least temporarily, a 1 WU per GPU policy and only until it either errors out, finishes or times out does it get another. This would dramatically speed up work flow | |
ID: 46317 | Rating: 0 | rate: / Reply Quote | |
You've made it perfectly clear. | |
ID: 46318 | Rating: 0 | rate: / Reply Quote | |
I actually think it will be just as effective when there are lots of WU's to be had because it's likely at least one or more types of work will have a higher priority setting than others so you don't want them cached in order to get them completed as fast as possible. | |
ID: 46319 | Rating: 0 | rate: / Reply Quote | |
I dont really second that as reducing the number of WU per GPU will lead to a poor utilization of capable Hardware (gtx 980, 980ti, 1070 and up) because of a bad CPU/GPU ratio. The 1 WU/GPU thing would only work well for mid range Maxwell Cards and most of the Kepler and Fermi Cards. But it will surely reduce the efficiency of the fast machines. | |
ID: 46329 | Rating: 0 | rate: / Reply Quote | |
the basic idea is not a bad one from various viewpoints; however, my reservations are mainly because of these facts brought up by Zoltan: ... If the backup project has long workunits, or the host has a slow internet connection (mainly hosts outside Europe) this reduction could be significant for these hosts (and for the whole project also). This would make some volunteers to disable their backup project(s), or manually manage them (which is pretty inconvenient). | |
ID: 46331 | Rating: 0 | rate: / Reply Quote | |
So perhaps this setting should be different for every host, based on the turnaround time of that host. Every host should start with 1 workunit per GPU, but if the turnaround time is below 48 hours, the maximum is set to 2. When the turnaround time is above 48 hours, the maximum is set to 1. Zoltan, that is a good proposal.. although I am not sure how to technically implement it, overriding the client config. ____________ I would love to see HCF1 protein folding and interaction simulations to help my little boy... someday. | |
ID: 46336 | Rating: 0 | rate: / Reply Quote | |
Thanks Betting Slip for the thread :) Very nicely put. | |
ID: 46346 | Rating: 0 | rate: / Reply Quote | |
Thanks Betting Slip for the thread :) Very nicely put. You're comments on the short queue are valid and would have put in my original post had I remembered, not to implement on short queue. I was so engrossed in making my case clearly that it slipped my mind. I think a 4 to 6 week trial would give both you and us a very good idea as to the impact on the project which, I personally think would be more positive than negative and most users would adapt their habits if running backup projects. As I said above, I also think it will be positive when there is lots of work. Be brave and give it a go. :-) | |
ID: 46347 | Rating: 0 | rate: / Reply Quote | |
May I put in one more remark.. the short run queue is intended for the slower cards where it doesn't make sense to run 2 concurrent jobs anyway. Whereas the fast Maxwell and Pascal cards need 2 tasks to have a good utilization otherwise they might be bottlenecked by the CPU. | |
ID: 46350 | Rating: 0 | rate: / Reply Quote | |
Sorry since I am not very familiar with BOINC. What is considered a host? One machine or one GPU? | |
ID: 46353 | Rating: 0 | rate: / Reply Quote | |
I would think it would be " Max WU in progress GPU" | |
ID: 46354 | Rating: 0 | rate: / Reply Quote | |
I am not exactly a Boinc expert either ... but can't the Boinc client check the estimated GFLOPS capability of the GPU used and send to the server? | |
ID: 46355 | Rating: 0 | rate: / Reply Quote | |
I would think it would be " Max WU in progress GPU" I'm not really an expert on practical server operations (just a volunteer user, like most people here), but I can find my way around well enough to point you to http://boinc.berkeley.edu/trac/wiki/ProjectOptions#Joblimits as a starting point. You probably need to read through the next section - Job limits (advanced) - as well. I'm guessing that because you operate with GPU limits already, there's probably a pre-existing config_aux.xml file on your server, which might be a better example to follow than the somewhat garbled one given as a template in the trac wiki. It appears that you can set limits both as totals for the project as a whole, and individually for each application (acemdlong / acemdshort). As Betting Slip says, a 'host' in BOINC-speak is one computer in its entirety - any number of CPU cores plus any number of GPUs (possibly of multiple technologies). I would guess that '<per_proc/>' shown for the <gpu_limit> in the wiki should be interpreted as 'per GPU' (bearing in mind that some dual-circuit-board GPUs, like the Titan Z, will appear to BOINC as two distinct GPUs), but that's an area where the documentation seems a little hazy. If you can clarify the situation by experimentation, I think I have an account which allows me to edit for clarity. | |
ID: 46356 | Rating: 0 | rate: / Reply Quote | |
If less than 2000 GFLOPS I would send only one short run per GPU. Otherwise grant 1-2 long runs. IF that is at all possible to configure. That seems pretty in-depth, I'm not sure it's that granular, but it's worth investigating. If not, I think the overarching 1 WU per GPU is the best route, at least for an experimental trial period. | |
ID: 46357 | Rating: 0 | rate: / Reply Quote | |
Sorry for spelling your name wrong Richard was going to go for the "s" but went with "z" instead. Does explane why I couldn't find you when doing a search, that should have given me a clue but no... | |
ID: 46358 | Rating: 0 | rate: / Reply Quote | |
Sorry for spelling your name wrong Richard was going to go for the "s" but went with "z" instead. Does explane why I couldn't find you when doing a search, that should have given me a clue but no... LOL - no offence taken. I'm used to the 'z' spelling when I dictate my surname over the telephone, but I'm always intrigued when somebody who only knows me from the written word plays it back in the alternate form. I guess the human brain vocalises as you read, and remembers the sounds better than the glyphs when the time comes to use it again? I'm sure there's a PhD in some form of cognitive psychology in there for someone... | |
ID: 46359 | Rating: 0 | rate: / Reply Quote | |
Will any server policy implemented include a fix for multiple architecture GPUs running simultaneously? | |
ID: 46362 | Rating: 0 | rate: / Reply Quote | |
I am running Kepler and Maxwell together so it must just be that they are different versions of CUDA, so I don't expect this to ever be fixed. But I could be wrong. | |
ID: 46363 | Rating: 0 | rate: / Reply Quote | |
At present there are not enough tasks to go around & most/all WU's have high GPU utilization so the project doesn't benefit (overall) from even high end GPU's running 2tasks at a time. Considering the recent upload/server availability issues it makes even less sense to run more than 1task/GPU. I would even go so far as to suggest its time we were running 1 task across multiple GPU's (for those who have them) - something that would also benefit those with 2 or more smaller GPU's; assuming an NV-link/SLi isn't required. By the way, GPU utilization isn't the end all be all of GPU usage; apps that perform more complex simulations, that require more GDDR, and more PCIE bandwidth are just using the GPU differently; not more or less (one probably comes at the expense of the other). | |
ID: 46381 | Rating: 0 | rate: / Reply Quote | |
At present there are not enough tasks to go around & most/all WU's have high GPU utilization so the project doesn't benefit (overall) from even high end GPU's running 2tasks at a time. I think though that Betting Slip meant his suggestion in a different way: not talking about 2 tasks running at a time, but rather one task running, with another task downloading already (long time) before the running task gets finished. However, I am asking whether the policy suggested by him has meanwhile been implemented: On one of my PCs, when the BOINC manager tried an Update on GPUGRID tasks (i.e. trying to download the next WU) while one WU was in progress, it said "the computer has reached a limit on tasks in progress". | |
ID: 46438 | Rating: 0 | rate: / Reply Quote | |
At present there are not enough tasks to go around & most/all WU's have high GPU utilization so the project doesn't benefit (overall) from even high end GPU's running 2tasks at a time. No, it has not been implemented evidenced by one of your computers with a single 750ti having 2 long WU's. How does that even begin to make any sense? http://www.gpugrid.net/results.php?hostid=372115 I'm not holding my breath on it being implemented either. | |
ID: 46439 | Rating: 0 | rate: / Reply Quote | |
No, it has not been implemented evidenced by one of your computers with a single 750ti having 2 long WU's. How does that even begin to make any sense? http://www.gpugrid.net/results.php?hostid=372115 Do you mind aborting at least one of those WUs? My two 1070s and 980ti are idling, and I'm sure many others are like me. | |
ID: 46440 | Rating: 0 | rate: / Reply Quote | |
Frankly, I don't think it's a good idea that crunchers are not starting to bite (not to say: attack) each other because there are WUs in waiting position at the host of a given cruncher, while the GPU of another cruncher is running idle. | |
ID: 46443 | Rating: 0 | rate: / Reply Quote | |
Do you mind aborting at least one of those WUs? My two 1070s and 980ti are idling, and I'm sure many others are like me.There's no point in aborting a task on an alive and active host, as there's nothing that could make sure that this task gets assigned to an idle and active host. As the number of tasks in progress decreases, there are more hosts trying to get work (in vain) including those which never finish the work. The overwhelmed / inactive hosts could impede the most the progress of a chain, as there's a 5 days deadline, so such a host could cause that much delay in a single step. When there's no work the chance of getting a previously timed out task is higher, even such tasks which timed out more times (causing 10, 15, 20... days delay). For example this one. It spent 10 days in vain before I've received it. | |
ID: 46468 | Rating: 0 | rate: / Reply Quote | |
Frankly, I don't think it's a good idea that crunchers are not starting to bite (not to say: attack) each other because there are WUs in waiting position at the host of a given cruncher, while the GPU of another cruncher is running idle. Right, squabbling will get us nowhere. Well ... seems that GPUGRID is self regulatory in terms of crunching power as many things in nature. I try not to take that personally. If there is not enough work, quite a few crunchers will walk away and seek for jobs elsewhere. I have already withdrawn my 1070 from GPUGRID and my remaining 1080 still is idle very often due to lacking tasks. Maybe I should withdraw this one as well. Leaves more work for others. ____________ I would love to see HCF1 protein folding and interaction simulations to help my little boy... someday. | |
ID: 46469 | Rating: 0 | rate: / Reply Quote | |
Zoltan you are very right, there is a risk, but I have been fast enough to claim every aborted task so far. And JoergF, there's no sense in completely withdrawing GPUs from GPUGrid, just have a backup with less priority so it automatically switches to that other work load if there is no work in GPUGrid. This is the only way I can keep my house warm with this intermittent work, and with no babysitting. | |
ID: 46471 | Rating: 0 | rate: / Reply Quote | |
Yes --- Leave GPUGrid attached! | |
ID: 46487 | Rating: 0 | rate: / Reply Quote | |
Somehow I manged to not get a single WU from the 600 new WUs released even though I manually updated all my my computers during the time they were still being claimed. | |
ID: 46488 | Rating: 0 | rate: / Reply Quote | |
Somehow I manged to not get a single WU from the 600 new WUs released even though I manually updated all my my computers during the time they were still being claimed.That's regrettable, but it's not unexpectable, and I see it as a justification of my observation I've posted earlier: "there's nothing that could make sure that this task gets assigned to an idle and active host." BTW, I haven't received from these workunits either. | |
ID: 46489 | Rating: 0 | rate: / Reply Quote | |
Has the 1 WU per GPU rule been enforced? It seems so. I have just purchased a new 1080ti and have very poor utilization (~70%) with only one 1WU (although the client config is set to 2WU in parallel). My 6700K (not OC yet) isn't fast enough to feed the 1080ti with data, so this is the outcome. My old 1080 had >90% with 2WU even under Windows 10. | |
ID: 47926 | Rating: 0 | rate: / Reply Quote | |
Well.. at present there are not many WU in the Queue anyway... :-( | |
ID: 47927 | Rating: 0 | rate: / Reply Quote | |
Has the 1 WU per GPU rule been enforced? It seems so. I have just purchased a new 1080ti and have very poor utilization (~70%) with only one 1WU (although the client config is set to 2WU in parallel). My 6700K (not OC yet) isn't fast enough to feed the 1080ti with data, so this is the outcome. My old 1080 had >90% with 2WU even under Windows 10. One WU per GPU has not been enforced, rather there just aren't enough WUs to go around so everyone is lucky if they get just one. As for the utilization, if you look at the Server Status link at the top right of the web page, you'll see that they run a fleet of different WUs, all with different utilization. Some with have 90%+ even on windows, and some, 70-80% on any machine and any operating system. There are ways to improve this. There is a variable called SWAN_SYNC in windows' environment variables. You get there by searching "variable" in the windows search. Once there you can click Environment variables on the bottom right. Under "user variables for your account" Click New and add SWAN_SYNC with a variable value of 1. Restart your computer and you will gain a few % in GPU utilization. | |
ID: 47928 | Rating: 0 | rate: / Reply Quote | |
I beg to differ. My app_config.xml is the same as always... but Boinc seems to ignore it. | |
ID: 47938 | Rating: 0 | rate: / Reply Quote | |
Have you tried disabling Hyper Threading (HT)? I've been experimenting and have seen slightly higher utilization when HT is off on my i7-7700k. CPU usage jumps from 13% to 25% per WU, and GPU-Z shows a small gain in usage. | |
ID: 47939 | Rating: 0 | rate: / Reply Quote | |
I've been experimenting and have seen slightly higher utilization when HT is off on my i7-7700k.That's because there are only 3 CPU tasks running when HT is off, and 7 when HT is on. You can gain a little further GPU utilization if you run only 1 CPU task, or no CPU tasks at all. [when HT is off] CPU usage jumps from 13% to 25% per WUThat's because when HT is on your CPU can handle 8 threads (on the 4 physical cores), so one thread uses 100%/8=12.5% (~13%), while your CPU can handle 4 threads on the 4 physical cores when HT is off, so one thread uses 100%/4=25% GPU-Z shows a small gain in usage [when HT is off].See my first explanation for its reason. (less CPU tasks = higher GPU usage) I wish there was a way to force the WU to use a full physical core without turning HT off, but I haven't been able to figure it out. I tried setting <cpu_usage>2.0</cpu_usage> with HT on, but that didn't help, still 13% usage.The <cpu_usage> option tells the BOINC manager how many cores (threads) to assign for the given application, it does not instructs the application to use that many cores (threads). You can use the SWAN_SYNC environmental variable to instruct the GPUGrid application to use a full core (thread), but it won't use more than one. My comments apply to Windows 10, cpu usage in Linux seems less, I'm not sure this would help there.There's no WDDM in Linux, so GPU usage will be higher under Linux than on Windows 10 (Vista, 7, 8, 8.1). | |
ID: 47940 | Rating: 0 | rate: / Reply Quote | |
less CPU tasks = higher GPU usage To clarify, are you saying that the only reason that disabling HT is showing an improvement is because it's reducing the number of CPU tasks from other projects? That would imply that the increase in CPU usage from 12.5% (HT) to 25% (HT off) does not in itself improve WU return times. For example, if no other projects were running, I would still think that more CPU usage is better regardless, but perhaps it is not that simple. | |
ID: 47941 | Rating: 0 | rate: / Reply Quote | |
Yes.less CPU tasks = higher GPU usageTo clarify, are you saying that the only reason that disabling HT is showing an improvement is because it's reducing the number of CPU tasks from other projects? That would imply that the increase in CPU usage from 12.5% (HT) to 25% (HT off) does not in itself improve WU return times.Yes. For example, if no other projects were running, I would still think that more CPU usage is better regardless, but perhaps it is not that simple.It's not that simple, as there are many factors which are reducing the GPU usage. WDDM has the most impact, and it can't be turned off; you should use Linux or Windows XP (up to GTX 980 Ti, so you can choose only Linux for your GTX 1080) More than 1 CPU task usually cause demonstrable decrease (depending on the task, as these are usually work on large data sets, which won't fit in the L3 cache of the CPU, thus they will flood the memory bus which will slow everything down a bit). Using SWAN_SYNC (in combination with no CPU tasks) could increase the the GPU usage on Windows XP by up to ~15%, and by up to ~5% on Windows 10. BTW you can try it very easily by simply suspending all CPU projects in BOINC manager for a test period. | |
ID: 47942 | Rating: 0 | rate: / Reply Quote | |
Retvari Zoltan wrote: Using SWAN_SYNC (in combination with no CPU tasks)... ..or just increase priority of GPU tasks. I wrote earlier in this message how to automate the process of increasing priority of GPU tasks. I see no reason to abandon computing CPU tasks, if there is a quite simple way to minimize their effect(or other programs, if it's not a dedicated host for computing) on GPU tasks. | |
ID: 47945 | Rating: 0 | rate: / Reply Quote | |
Aleksey Belkov wrote:
As people stated in your link... GPU tasks already run, by default, at a higher process priority (Below Normal process priority = 6) than CPU tasks (Idle process priority = 4). You can inspect the "Base Prio" column in Task Manager, or the "Priority" column in Process Explorer, to confirm. The Windows Task scheduler does a good job of honoring these priorities. So... Are you seeing a speedup by hacking the priorities with Process Hacker? And, if so, isn't the real reason that you see it is because you're bumping priorities to be higher than the (Normal priority = 8) tasks that Windows uses for all your other non-BOINC processes? BOINC is meant to run tasks without interfering with the normal operations of a device. So, I think our defaults work well to do that, and I think your hacking may work well to achieve more GPU throughput at the expense of potentially slowing down normal operations of the device which now run at a priority lower than your hacked processes. Regards, Jacob | |
ID: 47946 | Rating: 0 | rate: / Reply Quote | |
Or run #ofCPUThreads minus 2 to leave a full core for GPUGrid if you must. I sure wouldn't castrate CPU production from 7 to 3 tasks for a single GPU task. The 3 would run faster but production most likely would go down overall. There just needs to be more tasks. Plain and simple. There is a Formula BOINC competition right now for the next 3 days and it's not going to be a competition of who has the most processing power but who can get the most work. :( | |
ID: 47947 | Rating: 0 | rate: / Reply Quote | |
Jacob Klein wrote:
Probably I did not express myself correctly. Described method ensures that for GPU tasks will be allocated as much resources as they are requesting and CPU tasks will receive all remaining available resources(after execution of other processes having higher priority). In my opinion the method using SWAN_SYNC leads to loss of some CPU resources, which is rigidly assigned for GPU tasks. In previous tests I did not see a significant difference between using SWAN_SYNC and increasing priority, that has made me think that GPU tasks don't need such amount of CPU time. Therefore, it is sufficient to prioritise the execution of GPU tasks over other processes(not time-critical) to improve performance relative to the standard mode(without SWAN_SYNC and raising priority). I in my experience when priority of the GPU tasks increasing, I have not noticed any problems with responsiveness of system(in my case it isn't a dedicated host) or any other negative effects(unless you don't try to play demanding 3D games). Of course, on different systems and in different usage scenarios, the effect can vary greatly. I suggest you conduct your own tests on a dedicated host or home/work host. I believe that this method is particularly useful for those who run GPUGRID on home computers. | |
ID: 47948 | Rating: 0 | rate: / Reply Quote | |
Thanks for the reply. I'm curious how much of a speedup your process hacking actually gets? | |
ID: 47949 | Rating: 0 | rate: / Reply Quote | |
well ..... back to my question and the topic ... is the 1 Work Unit Per GPU rule now set? If so, may I protest. That measure has an enormous impact on my GPU utilization. | |
ID: 47950 | Rating: 0 | rate: / Reply Quote | |
No, I don't think it is set. What evidence do you have? I have a PC with 2 GPUs and 4 GPUGrid GPU tasks. | |
ID: 47951 | Rating: 0 | rate: / Reply Quote | |
is the 1 Work Unit Per GPU rule now set?It's not set. There was a debate about this earlier when there was a shortage, with minimal response from the staff. I think it would be much better if the low-end cards would be refused to get work from the long queue, as this is predictably futile considering the 5 days deadline. See this workunit. | |
ID: 47953 | Rating: 0 | rate: / Reply Quote | |
Thanks for the reply. I'm curious how much of a speedup your process hacking actually gets?I was experimenting with process priority and CPU affinity also on my various CPUs (i7-870, i7-980x, i7-4770k, i3-4160, i3-4360, i7-4930k), but the GPU performance gain from these tricks was negligible compared to the lack of WDDM, and less (or none) CPU tasks. This led me to the conclusion that a single PC could not excel simultaneously in GPU and CPU performance, thus there's no need for a high-end CPU (though it should be state of the art) to maximize high-end GPU performance. It is very hard to resist the temptation of running 12 CPU tasks on a very expensive (6-core+HT) CPU which will reduce GPU performance; thus it's better to build (at least) two separate PCs: one for CPU tasks and one for GPU tasks. This is the reason for my PCs with i3 CPUs: I rather spend more money on better GPUs than on better CPUs. Regarding RAC (or PPD): the loss of RAC of a high-end GPU by running CPU tasks simultaneously on the same PC is much bigger than the RAC gain of CPU tasks, so the overall RAC of the given PC will be lower if it runs CPU and GPU tasks simultaneously. Of course if someone could have only one PC, their possibilities for performance optimization are limited, and my advice is not worth to be applied (because of the huge loss in the user's overall CPU performance). Sorry for writing the same thing different ways many times, but my experiences are confirmed by the fact that my GTX 980Ti's (driven by i3 CPUs) are competing with GTX 1080Ti's on the performance page. Though the days of my GTX 980Ti's (running under Windows XPx64) are numbered, as the next generation (Volta) GPUs will wash them away from the top list even with WDDM, we could still use Linux to avoid WDDM, so my advice will still stand. | |
ID: 47954 | Rating: 0 | rate: / Reply Quote | |
Thanks Retvari. | |
ID: 47955 | Rating: 0 | rate: / Reply Quote | |
I can understand the different goals and mindsets with all the projects out there. I'd say my goals are more in line with Retvari's; fastest GPU WU return times possible. With this in mind, I've detached from all CPU projects and have seen improvements. The faster/ more GPU's you have in your system the more this makes sense. An improvement of only 1% on the return times from my pair of GTX 1080's easily provides more RAC than my i7-7700k could running CPU tasks alone. | |
ID: 47956 | Rating: 0 | rate: / Reply Quote | |
I would encourage you to do some testing, using the following BOINC client options (which you'd use in cc_config.xml), especially <process_priority_special>, which might allow you to do what you want without hacking. <no_priority_change>0|1</no_priority_change> <process_priority>N</process_priority> I bet they work for you!! Try them and let us know :) | |
ID: 47957 | Rating: 0 | rate: / Reply Quote | |
Try them and let us know :) I tried these changes to cc_config, and they worked for all of the other projects I tried, but not for GPUgrid. I was running CPU tasks on WCG and I run GPU tasks on Einstein when work is low here. I was able to manipulate the priority of each, setting a higher priority for the Einstein GPU tasks. No such luck on GPUgrid tasks; that's why I was so happy to hear about "Process Hacker". Maybe it's just me, has anyone successfully changed the priority of GPUgrid tasks with cc_config? | |
ID: 47958 | Rating: 0 | rate: / Reply Quote | |
I reproduced your problem - GPUGrid GPU tasks aren't honoring that <process_priority_special>. Other projects' GPU tasks do honor it. | |
ID: 47959 | Rating: 0 | rate: / Reply Quote | |
I reproduced your problem - GPUGrid GPU tasks aren't honoring that <process_priority_special>. Other projects' GPU tasks do honor it. Would swan_sync play a role with this priority option? | |
ID: 47960 | Rating: 0 | rate: / Reply Quote | |
Edit: Double post :( | |
ID: 47961 | Rating: 0 | rate: / Reply Quote | |
I reproduced your problem - GPUGrid GPU tasks aren't honoring that <process_priority_special>. Other projects' GPU tasks do honor it. No, I don't think so. | |
ID: 47962 | Rating: 0 | rate: / Reply Quote | |
mmonnin wrote:
This combination can lead to significant problems with responsiveness of system. | |
ID: 47963 | Rating: 0 | rate: / Reply Quote | |
I reproduced your problem - GPUGrid GPU tasks aren't honoring that <process_priority_special>. Other projects' GPU tasks do honor it.If I can recall it correctly, this behavior is intentional. Originally, the GPUGrid app run at the same process priority level ("Idle") as CPU tasks, but it turned out to hinder GPU performance. Back then this <process_priority_special> did not exist (or the staff thought that it wouldn't be used by many participants) so it's been hard coded into the app to run at "below normal" priority level. EDIT: it was the result of "iterating" the optimal process priority level, as when it was hard coded to "above normal" some systems (Core2 Duo and Core2 Quad using SWAN_SYNC) became sluggish back then. While it's not explicitly stated, see this post by GDF (and the whole thread). EDIT2: See this thread also. | |
ID: 47964 | Rating: 0 | rate: / Reply Quote | |
Well, it should be fixed. If the user is going to use Swan_Sync via manual system variable, they can use manual cc_config to control the priority (if the app isn't rude like it is currently). | |
ID: 47965 | Rating: 0 | rate: / Reply Quote | |
Message boards : Number crunching : 1 Work Unit Per GPU