Message boards : Number crunching : Credit per € / $
Author | Message |
---|---|
last update: 29th of April 2009 ExtraTerrestrial Apes wrote:
Would still be nice if someone could provide such performance numbers for other WUs. ____________ Scanning for our furry friends since Jan 2002 | |
ID: 8566 | Rating: 0 | rate: / Reply Quote | |
And some more points to consider: | |
ID: 8567 | Rating: 0 | rate: / Reply Quote | |
I cannot find the chart anymore but on it my 8800GTX is said to have a core G80 so not good for using GPU but.. | |
ID: 9401 | Rating: 0 | rate: / Reply Quote | |
During the first months (about 3?) G80 did run GPU-Grid with a separate code path, to work around the missing features. Later CUDA versions broke something. | |
ID: 9448 | Rating: 0 | rate: / Reply Quote | |
Aw ok thanks a lot. | |
ID: 9449 | Rating: 0 | rate: / Reply Quote | |
Hmm Not a thread about this. | |
ID: 9450 | Rating: 0 | rate: / Reply Quote | |
Not sure what you mean by maximising, but this is certainly the wrong thread for that. Generally there's nothing the control panel could do for CUDA. Maybe if your card runs GPU-Grid in 2D-Mode (check clocks with GPU-Z), but this is not generally the case (switches to 3D clocks automatically). | |
ID: 9523 | Rating: 0 | rate: / Reply Quote | |
Can we update this? I have a friend building a new PC and trying to figure out what card to get for part gaming, part crunching $100-300. | |
ID: 10621 | Rating: 0 | rate: / Reply Quote | |
Can we update this? I have a friend building a new PC and trying to figure out what card to get for part gaming, part crunching $100-300. What is needed to be updated? The only part that changes is the prices ... find the most productive card for the money he wants to spend. For that range any of the 200 series cards is possible. I have 260, 280 and 295 cards and there is not a huge difference in the throughput on these cards... though there is a detectable improvement as you go up in capacity ... any will be decent performers... | |
ID: 10626 | Rating: 0 | rate: / Reply Quote | |
Which is best bang for buck? 275 or 260 core 216? Or does it make sense to scale up to 285/295? Should he wait for GT300 architecture? | |
ID: 10629 | Rating: 0 | rate: / Reply Quote | |
We can't be sure about GT300 yet. It looks like an expensive monster.. certainly impressive, but bang for the buck we can not assess (yet). | |
ID: 10632 | Rating: 0 | rate: / Reply Quote | |
I found a couple posts where a person was saying their vid card could not meant the GPUGrid WU deadlines but he doesn't say what card that is. | |
ID: 11128 | Rating: 0 | rate: / Reply Quote | |
I found a couple posts where a person was saying their vid card could not meant the GPUGrid WU deadlines but he doesn't say what card that is. Yes, 9600GT can, but slow, 28 huors+ for a 93-GIANNI one.... <core_client_version>6.4.7</core_client_version> <![CDATA[ <stderr_txt> # Using CUDA device 0 # Device 0: "GeForce 9600 GT" # Clock rate: 1600000 kilohertz # Total amount of global memory: 536543232 bytes # Number of multiprocessors: 8 # Number of cores: 64 MDIO ERROR: cannot open file "restart.coor" # Time per step: 204.018 ms # Approximate elapsed time for entire WU: 102008.765 s called boinc_finish </stderr_txt> ]]> ____________ "Silakka" Hello from Turku > Åbo. | |
ID: 11129 | Rating: 0 | rate: / Reply Quote | |
Would still be nice if someone could provide such performance numbers for other WUs. Recently replaced a Palit GTS 250 with a Palit GTX 260 (216), so I have some performance numbers. Details & Specs: The GTX260 has two fans rather than the one on the GTS250. Although the GTX260 is louder, the temperature is a bit less; 71 °C (running GPUGrid) rather than 76°C. The GTX260 is a bit longer too, but both would make a BIOS reset awkward, so no messing. The 250 has VGA, DVI and HDMI, but is only 1.1 Compute Capable (CC); using the G92 core. The 1.3 CC (G200 core) GTX260 only has 2 DVI ports, but I have a DVI to HDMI converter and a DVI to VGA adapter, should I ever need them. Although the GTX260 has 27 Multiprocessors and 216 Shaders, compared to the GTS250’s 16 Multiprocessors and 128 Shaders, my system’s power usage is surprisingly similar, perhaps even slightly less for the GTX260! Surprising until I looked at the clock rates; GTS250 1.85GHz, GTX260 1.35GHz! http://www.techpowerup.com/gpuz/6y4yp/ http://www.techpowerup.com/gpuz/hkf67/ Apart from changing the cards, the system is identical and Granted credit was the same for both WU’s (5664.88715277777): On the GTS250 I completed the Work Unit 48-GIANNI_BINDTST001-7-100-RND7757_2 in 53268.082 s On the GTX260 I completed the Work Unit 334-GIANNI_BIND001-7-100-RND2726_1 in 31902.258 s The GTS250 has a Boinc GFlops rating of 84 while the GTX260 is 104, which would make the GTX almost 20% faster, going by the Boinc GFlops rating. However, the similar work unit did not complete in 80% of the time it took the GTS250 (which would have been 42614.4656 sec); it completed it in 59.89% of the time. So to turn that around the new card was between 40 and 41% faster overall, and about 20% faster than Boinc predicted with its GFlops rating (Explained by having the better 1.3CC G200 core, compared to the 1.1CC G92 core of the GTS250). So for the above conditions, I would say the GTX260 has a rating of 104 (125) Boinc GFlops, with the number in the brackets representing the 1.3 CC/G200 comparable value (or a 1.2 correction factor; +20%) to a 1.1CC G92 core, that was rated at 104 Boinc GFlops. Perhaps these results vary with different tasks and cards? ____________ FAQ's HOW TO: - Opt out of Beta Tests - Ask for Help | |
ID: 12196 | Rating: 0 | rate: / Reply Quote | |
Thanks for reporting. Your numbers actually exactly confirm what I found, you just got it wrong at the end ;) | |
ID: 12206 | Rating: 0 | rate: / Reply Quote | |
It the math is done, it's a question of getting the max credit per $/€ invested + the cost of powering the beast. I'm not in this for the credits, nor do I have any special interest, other than hoping that I can help the collective, because I might need some helping out myself someday. | |
ID: 14814 | Rating: 0 | rate: / Reply Quote | |
last update: 29th of April 2009 Efficiency & total potential is a factor, but what about initial cost of purchase vs running cost? That prices swing & that some can be lucky as to get a good deal on a GPU, & that the price of power is different from country to country. That some of use use a 85+ PSU & if it's winter & the beast warms up the room, is also a factor. ____________ | |
ID: 14819 | Rating: 0 | rate: / Reply Quote | |
There are lots of factors to consider. Initial purchase costs, running costs, actual contribution (points being a good indicator), reliability, even aesthetics perhaps, certainly noise, heat and other unwanted/wanted side effects! | |
ID: 14844 | Rating: 0 | rate: / Reply Quote | |
24/7 operation causing a shortening of life span, although that 3 years will not be an issue, if it's credits you'd want to generate. Mine's factory OC'd plus slightly more. That 1 PC running several GPU's is overall more efficient than several PC's running just one GPU. | |
ID: 14848 | Rating: 0 | rate: / Reply Quote | |
Just moved the GT 240 to an i7, to see how much it benefits from the added CPU power. I have not tried to OC it yet, as I was mainly concerned with reliability. I did underclock the opteron from 2.2GHz down to about 800MHz and the time to complete tasks dropped significantly for the GPU, so giving it faster CPU support should reduce task turnover times. Its low temperatures should allow it to clock reasonable well. I will try to OC it in a few days, after I have a better i7 heatsink, as the i7 is running too hot and could interfere with the overclocking the GPU. Its also good to get some stock results in first. | |
ID: 14907 | Rating: 0 | rate: / Reply Quote | |
Is the GT240 still the best bang for the buck? I'm looking at a 9800GT "Green", which has 112 Stream Processors, a 550Mhz Core Clock, no PCIe power connector (like the GT240). But I'm also looking at the GTS240 which I "suspect", is just a 9800GT "Rebranded Non-Green AKA Original 8800GT/9800GT" version of the 9800GT "Green". The GTS240 also has the missing PCIe power connector & a 675Mhz Core Clock, instead of the "Green" 550Mhz Core Clock. BTW, what makes most sense? The GT240 with 96 Stream Processors a 550Mhz Core Clock - with the potential of a higher OC, or the 9800GT "Green" - with 112 Stream Processors - but a possible lower OC potential??? | |
ID: 15704 | Rating: 0 | rate: / Reply Quote | |
I will say that the number of SP more important than the freq. (Obviously we have to check both) but in such a difference, I will prefer a 9800GT. | |
ID: 15712 | Rating: 0 | rate: / Reply Quote | |
I will say that the number of SP more important than the freq. (Obviously we have to check both) but in such a difference, I will prefer a 9800GT. But I do like the G200, but I can find a single one that I can fit into a single slot nano-BTX casing, nor the micro-ATX HTPC casing. I've got 3 GTX260-216, & I do love them! Even more then the GTX275, maybe because I'm so cheap. ____________ | |
ID: 15715 | Rating: 0 | rate: / Reply Quote | |
Guys.. | |
ID: 17592 | Rating: 0 | rate: / Reply Quote | |
Guys.. Now that Nvidia is so good at rebranding, reusing, & slightly changing old chips. Would there be any sense in a 40nm version of the Asus Mars 295 Limited Edition? http://www.techpowerup.com/95445/ASUS_Designs_Own_Monster_Dual-GTX_285_4_GB_Graphics_Card.html It's old-tech, yes, but so is the GTS250/9800GTX/8800GTX. Maybe a 32nm flavor with GDDR5? If a 32nm is something Nvidia "might" want to do, would it be small enough, & possible to "glue" it together like what Intel did with their Core2 Quad? Two steps forward & one step back, just like Intel with PIII & P4... ____________ | |
ID: 17593 | Rating: 0 | rate: / Reply Quote | |
They should be shipping the improved 40 nm version of GT200 in less than a month. It's called GF104 and is even better than a shrink of GT200: it's a significantly improved, more flexible and more efficient design due to its Fermi heritage. Remember: GF100 is not very attractive because it's insanely large and because it doesn't have enough TMUs. A half GF100 + double the amount of TMUs can fix both. See the Fermi thread in the other sub forum :) | |
ID: 17595 | Rating: 0 | rate: / Reply Quote | |
This is worth a quick look, | |
ID: 18333 | Rating: 0 | rate: / Reply Quote | |
One of the calculations that needs to be included in a credit per dollar table is the frequency in GPUGrid of medium to long run computation errors which GPUGrid work units are very much prone to. This has been something of a long running problems (months, perhaps years). | |
ID: 19025 | Rating: 0 | rate: / Reply Quote | |
I was running 4x 9800GT (G92 rev A2) on 4 different system with WinXP-32bit, Win Vista 32bit/64bit and Win7 64bit. | |
ID: 19026 | Rating: 0 | rate: / Reply Quote | |
Well, I finished a couple of those LONG tasks and good grief they are REALLY long. | |
ID: 21238 | Rating: 0 | rate: / Reply Quote | |
Different task types take different amounts of time to complete. | |
ID: 21242 | Rating: 0 | rate: / Reply Quote | |
The EVGA GTX 560 Ti's versions I have are the Super Clocked varient and run 900MHZ normally and that's where they sit burning in before I install waterblocks on them. They pass all the hardest 3D tests and Memory tests I can throw at them, so no problems there I think. | |
ID: 21244 | Rating: 0 | rate: / Reply Quote | |
This under-clocking is becoming one of the biggest self-inflicted problems I have seen in GPU computing - an untreated plague. Only equaled by the unusable shaders (2 warp schedulers per 3 shader groups), and heat/fan issues of present/recent past - avoidable design/driver flaws. | |
ID: 21249 | Rating: 0 | rate: / Reply Quote | |
This under-clocking is becoming one of the biggest self-inflicted problems I have seen in GPU computing - an untreated plague. Only equaled by the unusable shaders (2 warp schedulers per 3 shader groups), and heat/fan issues of present/recent past - avoidable design/driver flaws. A quite reliable solution I found is setting power management from adaptive to maximum performance. This way in case of computation errors the clocks won't fall back to 2D level and you can increase them again using Afterburner without a reboot. What I found as most frequent reason for this downclocking - at least on Primegrid - is not enough voltage causing the computation error first. In my few efforts here GPUGRID was even more sensitive and hardly could run the same clocks. ____________ | |
ID: 21584 | Rating: 0 | rate: / Reply Quote | |
Unfortunately, under XP you can no longer set the GPU to Maximum Performance; it's stuck at Adaptive. | |
ID: 21606 | Rating: 0 | rate: / Reply Quote | |
I hope it's the right thread - I'm mostly trying to optimise Credit/Watt on that next machine i'm building. Initial price is a minor issue. Obviously an efficient PSU and SSD will play a big role for the system. My question here: | |
ID: 21898 | Rating: 0 | rate: / Reply Quote | |
Sandy Bridge processors are more energy efficient than previous generations by Intel; a stock i7-2600 uses around 65W when crunching 8 CPU projects, an i7-920 (previous generation) uses over 100W and does less work. In terms of CPU crunching per Watt an i7-2600 is about 50% more efficient than an i7-920. | |
ID: 21902 | Rating: 0 | rate: / Reply Quote | |
Thanks, skgiven. Very informative answer! | |
ID: 21930 | Rating: 0 | rate: / Reply Quote | |
The Intel Core i7-970, i7-980, i7-980X, i7-990X six core socket 1366 CPUs are also made on 32nm technology, without on-die GPU. You should consider them also, especially if you want to crunch with two GPUs. While the X58 chipset for socket 1366 CPUs have two native 16x PCIe 2.0 connectors, Socket 1156 and Socket 1155 CPUs have only one 16x PCIe 2.0 bus integrated into the CPU. So if someone uses two GPUs with these CPUs, the GPUs will have only 8x PCIe 2.0 per GPU, and it will lower the performance of the GPUGrid client. However, ASUS made a special MB for socket 1156 CPUs, with two native 16x PCIe 2.0 connectors: the P7P55D WS Supercomputer, and for socket 1155 CPUs (SandyBridge): P8P67 WS Revolution. Maybe other manufacturers have similar motherboard designs. | |
ID: 21932 | Rating: 0 | rate: / Reply Quote | |
I still don't think X8 makes all that much difference; well not in itself for 2 cards. For two GTX590's perhaps. There are other factors such as the 1366 boards having triple channel memory, 36 PCIE lanes, and 1156/1155 only having dual channel memory. The chipset that controls the PCIE lanes also controls USB, PCI, SATA and LAN, so what else is going on might influence performance. So it's really down to the motherboards implementation and what the system is being used for. Not sure about that P8P67 WS Revolution LGA 1155 Motherboard (2 x PCIe 2.0 x16 (x16, x8)), but the GA-Z68X-UD7-B3 is a bespoke implementation of an LGA 1155 motherboard that offers two full PCIE X16 lanes (x16, x16). | |
ID: 22055 | Rating: 0 | rate: / Reply Quote | |
I still don't think X8 makes all that much difference; well not in itself for 2 cards. For two GTX590's perhaps. I have 4 GTX 480s in two PCs: 1. Core2 Quad 9650 @4GHz (FSB 444MHz), X48 chipset, dual native PCIe x16 slots. 2. Core i7-870 @3.92GHz (BCLK 180), P55 chipset (PCIe x16 controller in the CPU) After overclocking the Core 2 Quad, it is faster (got higher RAC) than the Core i7-870, either I put the 2 GTX 480s in the x16 slots (and have only x8 for both), or I put the second card to the third PCIe slot (and have x16 for the first card, and x4 for the second). The lower the GPU usage of a WU, the higher the impact on the performance of slower PCIe. As far as I can recall, it was around 10% comparing x8 and x16. You can see the difference (it's now around 20%) of x4 and x16 among my host's reults. Anyway when the Sandy Bridge E type CPU's turn up their boards will have support Quad Channel Memory, PCI-e 3.0 and multiple x16 PCI-e lanes. Obviously these will become the CPU/board combination to have. Of course, but I have to add that they will be quite expensive. | |
ID: 22056 | Rating: 0 | rate: / Reply Quote | |
I see your Core2 Quad cards are about 12% slower than your i7-870's X16 GPU, but your i7-870's X16 GPU is 17% faster than the X4 GPU (just going by a few tasks). So overall your C2Q is only around 3.5% slower. Not sure if you are using HT or not, and what else you are crunching? HT might make a slight difference between your CPU types, C2Q and i7-870, but only a little if any between X16, X8 and X4. What the CPU is being used for and what else the controller is spending time on could also make a difference. The faster the CPU and GPU, the more it would potentially expose the weakness in PCIE lanes. | |
ID: 22058 | Rating: 0 | rate: / Reply Quote | |
I'm still satisfied with my 2 year 'old' C2Extreme X9650, now running @ 3.51GHz, | |
ID: 22116 | Rating: 0 | rate: / Reply Quote | |
Not sure if you are using HT or not, and what else you are crunching? I'm using HT, but only 4 tasks are running at the same time. (2 GPUGrid and 2 Rosetta@home) I expect the difference might be different for different cards. PCI x8 deficiency might be slightly less for a GTX 460 and cards such as the GTS 450, but more for a GTX580. It's obvious. That's why the GTX 590 (made of 2 underclocked GTX 580s) has its own NForce 200 PCIe x16 to 2 PCIe x16 bridge. In this way both chips have PCIe x16. Could NVidia put 2 GPUs on a single board without this bridge chip, and both GPUs would have PCIe x8 then, but it would decrease the the overall performance too much. It's impractical for a top-end dual GPU card. Would be interesting to know what this is for other cards in practice under different architectures (LGA 1366, 1155, AMD rigs) as well as your LGA 1156 setup. I couldn't resist to have the ASUS P7P55 WS SuperComputer MB I've mentioned earlier, so now my LGA1156 host runs with it. I kept the old P7P55D Deluxe MB too (and I bought an other Core i7-870 for it), so I can put the two GTX 480s from my Core2 Quad host to this MB, and we can compare the dual PCIe x8 setup with the dual PCIe x16 setup. | |
ID: 22160 | Rating: 0 | rate: / Reply Quote | |
Tested one of my i7-2600 systems, with a GTX470 @657MHz. Turns out the Gigabyte PH67A-UD3-B3 has one x16 slot and one x4 slot! The difference was around 13%. So a GTX580 would be >13% slower and a GTX460 would be less. Looked at several 1155 motherboards and on some the second slot is only x1 when both are occupied. | |
ID: 22200 | Rating: 0 | rate: / Reply Quote | |
PCIE 3 is also now available on some motherboards (MSI mostly). While these boards still tend to offer one X16 and either one x8 or x4 the improvement from PCIE2.0 is basically two fold; the bandwidth doubles from PCIE 2 to PCIE 3. | |
ID: 22210 | Rating: 0 | rate: / Reply Quote | |
I'm thinking about Gigabyte GA-890FXA-UD7 and 3x570. | |
ID: 22441 | Rating: 0 | rate: / Reply Quote | |
A guesstimate, but for a drop from PCIE x16 to x8 you might see a performance drop of around 8% on each GTX570. A further PCIE bandwidth drop to X4 would see the performance drop to around 16% less than X16. So overall you would be losing about 1/3rd of a GPU. | |
ID: 22447 | Rating: 0 | rate: / Reply Quote | |
Since we're talking about PCIe lately, I've also got a question | |
ID: 22631 | Rating: 0 | rate: / Reply Quote | |
My guess is between 4 and 8%, but let us know, when you get your motherboard and run a few tasks. | |
ID: 22632 | Rating: 0 | rate: / Reply Quote | |
Looks like the nvidia 680 690 are the way to go. | |
ID: 27132 | Rating: 0 | rate: / Reply Quote | |
Going by wiki, a GTX660Ti and a GTX660 are better than a GTX680 and a GTX670!
GTX660Ti 16.40GFlops/Watt GTX670 14.47GFlops/Watt GTX680 15.85GFlops/Watt GTX690 18.74GFlops/Watt
| |
ID: 27141 | Rating: 0 | rate: / Reply Quote | |
My guess is between 4 and 8%, but let us know, when you get your motherboard and run a few tasks. I have seen webpages where people also say the difference is not as great as one would think it might be. 10% is not that big of a deal when talking about a 150.00US motherboard. Now going to a new mb and going to 16gb of ram too IS a big deal! The ram of course is not used for gpu crunching very much, but does help when crunching cpu units too. Getting a new mb, new ram AND a new cpu along with that pcie-2 slot...now THAT is DEFINITELY worth it!! | |
ID: 27460 | Rating: 0 | rate: / Reply Quote | |
Anyone up to updating the OP? | |
ID: 28874 | Rating: 0 | rate: / Reply Quote | |
That's not an issue any more. The 4.2 apps do a good job of using the cc2.1 Fermis. MJH | |
ID: 28877 | Rating: 0 | rate: / Reply Quote | |
If only 2/3 of the shaders are being used why are the temps relatively high as is the GP utilization? Thanks for the update, I thought something had changed for the better. | |
ID: 28882 | Rating: 0 | rate: / Reply Quote | |
Anyone up to updating the OP? Interesting... I notice that the GIANNI (short tasks) uses just under 84% cpu! Not only that, it takes 124k-sec for a wu!!! That is almost 2x it takes to do the NATHAN (long tasks), which average 72k-sec on my Quadro-K2000M (Keppler w/2GB RAM). Can anyone explain why the 'short' tasks take almost 2x as long as the 'long' versions? My next step, out of sheer curiosity, I will try the same tests with my other machine which is a similar video, but Fermi design (Quadro-2000M). Thank you in advance. Phil ____________ | |
ID: 30136 | Rating: 0 | rate: / Reply Quote | |
http://www.gpugrid.net/forum_thread.php?id=3370&nowrap=true#30017 | |
ID: 30144 | Rating: 0 | rate: / Reply Quote | |
http://www.gpugrid.net/forum_thread.php?id=3370&nowrap=true#30017 Uh, yeah. Hello. Here is a copy of the linked 'response' you posted: There are two main reasons why more credit is awarded on the long queue: Very neat-o. This explains how GPUGrid came up with their terminology.... it does NOT answer my question. Using the same equipment, these GIANNI are taking 2x as long to run as the NATHAN. Why? You are supposedly compiling with the same options and both are CUDA42. Can somebody answer this question that has more than 15 posts experience; someone who will read the question before attempting to provide a single word answer? Thank you. ____________ | |
ID: 30193 | Rating: 0 | rate: / Reply Quote | |
Gianni's WUs may be slightly long for the short queue, but my recent tasks were definitely short for the long queue Other than that, if you are asking why system A is slower than system B then the answer is probably that system A contains more atoms / is more complicated, or the person sending the simulations asked for more simulation steps. Not all simulations are born equal :) I think the new NATHAN's are indeed long enough for the long queue. I hope that answered your question. I just wanted to point out that the NATHAN's you were talking about were not "standard" for the long queue. If you suspect it's your hardware setup then I unfortunately cannot help you and maybe you should make a new thread about it for someone to help you with. Cheers | |
ID: 30195 | Rating: 0 | rate: / Reply Quote | |
Back to Credit per €: | |
ID: 37798 | Rating: 0 | rate: / Reply Quote | |
More ideas for a efficient use requested ;-) Four GTX970's in a quad core system that supports 4 PCIE slots would probably be the most economically productive system at present. You could probably build a base unit for < £1500, get 2.5M credits/day (at present performance rates) and over 2 years at £0.15/Kw it would cost ~£1500 to run (assuming 582W draw) at which time the value of the system would be ~£500. So total cost of ownership (TCO) would be ~£2500 (~$4000 or ~3200Euro) over 2 years for a credit of 1.8B. Obviously prices vary by location and running costs depend on the price of electric which varies greatly from one place to another and the app could improve performance... For comparison, a quad GTX780Ti system would yield ~3.3M credits per day, but the purchase cost would be >£2100, the running cost would be ~£2600 over 2years and the system would only be worth about the same (older kit). So £4200 for 2.4B. 9700 system ~750M credits/£ over 2years 780 Ti rig ~570M credits/£ over 2years ____________ FAQ's HOW TO: - Opt out of Beta Tests - Ask for Help | |
ID: 38135 | Rating: 0 | rate: / Reply Quote | |
More ideas for a efficient use requested ;-) Skgiven, or anybody else--- I wondering what you would recommend for GPUGRID? I'm in the process of Building a DIY Haswell Xeon ( LGA-2011-3 socket for two CPU on MB/Quad channel memory) and 2 Low powered Z97 Dual channel systems. Two boards will have as many GPU possible. Also a decent amount of storage. ( A third will have 2 GTX660ti and maybe GTX750ti included.) My electric rates sky rocketed over few months, where I currently reside. (It has it own power station, and Town power company, but there been ongoing dispute about cost of distribution, creating rates from 0.0876cents/Kw to over 0.1768cents/Kw. These prices will be in effect for at least 6-9 months unless term agreement can be meant on both sides. Also, State energy taxes rose 20% since September of last year. For my LGA-2011-3 (2P) Quad channel DDR4 MB I will either choose a ASUS Z10PE-D8 WS (600usd) or GIGABYTE GA-7PESH3[640usd] (or something else), possible to have more than 4 GPU's on GA motherboard. For processors (quad channel), I already bought [2]-85TDP 6c/6t 2603V3 for 220usd each. I'm also considering a Z97 Board (LGA1150) for a low powered Xeon or i5/i7. (25TDP- 1240LV3 (4c8t) listed at 278$) or a 192usd i5-4590T[4c4t]45WTDP or a 303usd 35TDP i7 4785T[4c8t]. All newly bought GPU's will be Maxwell based. I'm thinking for LGA-2011-3 board, [4] GTX 970, (GTX980 price/performance/Wattage ratio is higher as you say.) unless a flaw been discovered in GPC. For the Z97 board, if a GTX 960 released shortly, I'll get [3]. I was given a couple second hand GTX660ti, by my sister, when she updated to GTX 980 last week for the gaming system I built for her. ( she lives out of state, so I don't access to test new GM204. So for these Two C.C3.0 cards I was thinking about picking up Z97 board, and a i5 4590T. I will "eco-tune" all cards including Maxwell, while waiting for GTX960 prices. (It's possible I keep GTX660ti, when GTX960 is released, or get more GTX970, depending on GTX960 price.) Total system(s) cost (5 GTX 970 or 3 GTX 970/2 GTX 960/ 1 GTX750ti/ 2 Xeon 2603V3/ 2 i5 4590T or 2 Xeon 1240LV3/1 Server MB/ 2 Z97 MB) will be (not including) PSU units will be around 3680- 4000usd, which is being piece meal ( looking for all and any discounts.) Thank you for help and advice. | |
ID: 38162 | Rating: 0 | rate: / Reply Quote | |
Four GTX970's in a quad core system that supports 4 PCIE slots would probably be the most economically productive system at present. Good Idea.. Is it necessary that all PCIe slots are running with x16 to obtain the full performance of the cards? Or is full power also available with 4 x8? creating rates from 0.0876cents/Kw to over 0.1768cents/Kw. A perfect and more than affordable energy world. Here in Germany we can only dream of prices like 0.1768 cents/kwh. Current price I have to pay is 0.27 Euro/kw/h which is 0.3406401 US-Dollar. Is it true? You pay only 0.1768 cents/kwh? ____________ Regards, Josef | |
ID: 38236 | Rating: 0 | rate: / Reply Quote | |
WE use the Asus Z87-WS motherboard - 4 GPUs at 8x. | |
ID: 38237 | Rating: 0 | rate: / Reply Quote | |
I made a small mistake figuring kwh cost. Considering how energy costs are all over place, and each Country applies a different formula for prices- I went back to look at my 2014 August-September Power bill to see exactly what is included-- currently 0.1112c/kwh is the flat rate (was 0.08-0.10c/kwh from Sept2013-March2014), add a 0.05982c/kwh surcharge for taxes, and 0.0354c/kwh for distribution. This beings it currently to ~0.19cents/kwh/usd. Natural Gas (my hot water) is opposite: taxes are less than distribution costs. I have three wood stoves for heat during winter. (Electrical costs are decent compared to 0.34c/kwh for Germany) In 2013- average total for kwh was 0.1445c. Average amount of "surcharges" in 2013 for each kwh was 0.0443cents. Cost for energy resources have rose considerably in every industrial area. | |
ID: 38238 | Rating: 0 | rate: / Reply Quote | |
I the course of upgrading my GPUs to Pascal I have worked through a lot of specs and characteristics. I thought that Maxwell was already a major leap forward in regard to efficiency and was somewhat surprised when I compared multiple generations by the GFLOP to power ratio. (Power draw taken from Wikipedia, typical 3D consumption) See the graph below. | |
ID: 45122 | Rating: 0 | rate: / Reply Quote | |
Nice graph, and I completely agree with what you said about upgrading! Kepler apparently performs better than its reputation. Maxwell (to my surprise) improved the efficiency just moderately. It's because they are based on the same 28nm lithography (TSMC had to skip the 20nm step, proposed to use for Maxwell). | |
ID: 45123 | Rating: 0 | rate: / Reply Quote | |
I the course of upgrading my GPUs to Pascal I have worked through a lot of specs and characteristics. I thought that Maxwell was already a major leap forward in regard to efficiency and was somewhat surprised when I compared multiple generations by the GFLOP to power ratio. (Power draw taken from Wikipedia, typical 3D consumption) See the graph below. Really? What about price? Not everyone, should I say most people can afford new Pascal cards. So, are they now regarded as secomd class. I also woud like to add that most Pascal card owners are a part of a "vanity" project. | |
ID: 45124 | Rating: 0 | rate: / Reply Quote | |
Really? What about price? Not everyone, should I say most people can afford new Pascal cards. So, are they now regarded as second class. I consider the older cards as second class, as being that much more effective than the previous generation(s) the Pascal's price can be saved in a short term on electricity costs. I also would like to add that most Pascal card owners are a part of a "vanity" project. I would rather call it "green" project. (Not because it's NVidia's color, but because it's twice as environment friendly as older cards) | |
ID: 45125 | Rating: 0 | rate: / Reply Quote | |
No offense meant. The below graph just illustrates very clearly that Pascal GPUs pay off quickly for hardcore crunchers. For example the GTX 780ti is still a well performing high end Card, but the GTX 1070 yields the same GFLOPs at 70-80W less. If you run the GPU almost day and night you will have extra cost of 0.08kW*8000hrs*0,2€/kWh=120€. A new GTX 1070 is available from 400€ and the GTX 780ti can be sold as second hand for maybe 150€. Which means you get a new card without further investment after two years, just because of saving energy. | |
ID: 45126 | Rating: 0 | rate: / Reply Quote | |
Other examples. Take the renowned Kepler GTX 780. The new GTX 1060 yields 10% more SP GFLOPs at 80-100W less power consumption. Leads to the same saving of 100-150€ per year. I would rather call it "green" project. (Not because it's NVidia's color, but because it's twice as environment friendly as older cards) yes, that is another valid argument. Really? What about price? Not everyone, should I say most people can afford new Pascal cards. So, are they now regarded as secomd class. I also woud like to add that most Pascal card owners are a part of a "vanity" project. I would neither label Maxwell users as "second class" nor Pascal users as "Snobs". If someone aims at a well performing card and just crunches now and again, a GTX 980 (maybe aside from the low 4GB Memory size) is absolutely a good choice. But for 24/7 operation Pascal makes much more sense in terms of energy cost. ____________ I would love to see HCF1 protein folding and interaction simulations to help my little boy... someday. | |
ID: 45129 | Rating: 0 | rate: / Reply Quote | |
No offense meant. The below graph just illustrates very clearly that Pascal GPUs pay off quickly for hardcore crunchers. For example the GTX 780ti is still a well performing high end Card, but the GTX 1070 yields the same GFLOPs at 70-80W less. If you run the GPU almost day and night you will have extra cost of 0.08kW*8000hrs*0,2€/kWh=120€ ...I have meant annual extra cost of course. Your electricity rate (€ per kWh) may differ from my above example. ____________ I would love to see HCF1 protein folding and interaction simulations to help my little boy... someday. | |
ID: 45136 | Rating: 0 | rate: / Reply Quote | |
That chart might not accurately reflect performance running GPUGrid apps! GeForce GTX TFlops (Boost) @2GHz £UKCost GFlops/£Cost
1050 2.56 120 21.3
1050Ti 3.07 140 21.9
1060 3GB 4.61 190 24.3
1060 6GB 5.12 240 21.3
1070 7.68 395 19.4
1080 10.24 600 17.1 Assumes all cards boost to 2GHz (and the 14nm cards might not). This is only theoretical and ignores factors such as app scaling, supporting OS, CPU, RAM, CPU cache, GPU L2 cache, task performance variations… Both performance/Watt and performance/purchase cost (outlay) are relevant. Reports of 45W actual use when crunching here on a GTX1060-3GB might throw the previous graphs data all over the place. The best measurements are actual observations (for here) and not theoretical. So what do cards actually boost to & what is the actual performance running specific task types? (post elsewhere or adapt for here). ____________ FAQ's HOW TO: - Opt out of Beta Tests - Ask for Help | |
ID: 45140 | Rating: 0 | rate: / Reply Quote | |
Taking performance/purchase cost into consideration is a valid argument. But in this case also the service life until replacing the GPU is of importance in order to calculate the amortization. Reports of 45W actual use when crunching here on a GTX1060-3GB might throw the previous graphs data all over the place. Well, I have already read many different power draw and utilization statements in this forum affecting all kind of GPU. From 45W to 65W or more for the 1060 in particular. So I guess we will never have perfect numbers and utilization of cards under every OS. From that view, the below graph should be a workable reference point, although far from perfection. ____________ I would love to see HCF1 protein folding and interaction simulations to help my little boy... someday. | |
ID: 45197 | Rating: 0 | rate: / Reply Quote | |
IMO 2years is a reasonable 'working' life expectancy for mid to high end GPU's, and that should form part of any TCO analysis for system builds/refurbs/performance per cost measurements. | |
ID: 45213 | Rating: 0 | rate: / Reply Quote | |
Does anyone know what is the better choice - 750 Ti or 1050Ti (both 4Gb)? | |
ID: 45255 | Rating: 0 | rate: / Reply Quote | |
Does anyone know what is the better choice - 750 Ti or 1050Ti (both 4Gb)? 1050 Ti of course. | |
ID: 45257 | Rating: 0 | rate: / Reply Quote | |
As an addition to my below graph, the new 1080ti yields 53 GFLOPS/Watt and the Titan XP about the same, assuming that it pulls 220-230W just like the Titan X. Which means the efficiency is identically equal to the non-"ti" 1080. So much for theory... | |
ID: 47018 | Rating: 0 | rate: / Reply Quote | |
As an addition to my below graph, the new 1080ti yields 53 GFLOPS/Watt and the Titan XP about the same, assuming that it pulls 220-230W just like the Titan X. Which means the efficiency is identically equal to the non-"ti" 1080. So much for theory... Since the 1070 is the same GP104 die, it performs very similarly to the gtx 1080 in this project, costs much less, and uses slightly less power. The 1070 is the best performance per watt, performance per dollar card for this project. | |
ID: 47110 | Rating: 0 | rate: / Reply Quote | |
As an addition to my below graph, the new 1080ti yields 53 GFLOPS/Watt and the Titan XP about the same, assuming that it pulls 220-230W just like the Titan X. Which means the efficiency is identically equal to the non-"ti" 1080. So much for theory... I agree. ____________ I would love to see HCF1 protein folding and interaction simulations to help my little boy... someday. | |
ID: 47111 | Rating: 0 | rate: / Reply Quote | |
Since the 1070 is the same GP104 die, it performs very similarly to the gtx 1080 in this project, costs much less, and uses slightly less power. The 1070 is the best performance per watt, performance per dollar card for this project. I have a 1660Ti that I would like to test out but but but … 1487 GPUGRID 8/15/2019 10:11:04 AM No tasks are available for Short runs (2-3 hours on fastest card) 1488 GPUGRID 8/15/2019 10:11:04 AM No tasks are available for Long runs (8-12 hours on fastest card) 1489 GPUGRID 8/15/2019 10:11:04 AM No tasks are available for New version of ACEMD 1490 GPUGRID 8/15/2019 10:11:04 AM No tasks are available for Anaconda Python 3 Environment 1491 GPUGRID 8/15/2019 10:11:04 AM Project has no tasks available | |
ID: 52513 | Rating: 0 | rate: / Reply Quote | |
well, you have been with us for 11 years now; so you should have gotten used to this kind of situation :-) | |
ID: 52514 | Rating: 0 | rate: / Reply Quote | |
I am currently running a GTX 650 (from an older pc) and a GTX 1050TI. | |
ID: 53064 | Rating: 0 | rate: / Reply Quote | |
Running a 1060 and quite happy with it. | |
ID: 53067 | Rating: 0 | rate: / Reply Quote | |
So which is better for the money spent and the power used? Following thread might be of your interest: http://www.gpugrid.net/forum_thread.php?id=4987 Personally, I find GTX 1650 interesting due to its relative low cost, power consumption and moderate performance. I have one of this currently working, processing about 2,5 ACEMD3 test tasks per day. And I See your 1050 Ti is currently failing all tasks. I had the same problem with the same card model. How the problem was solved can be found at this other thread: http://www.gpugrid.net/forum_thread.php?id=4999 | |
ID: 53077 | Rating: 0 | rate: / Reply Quote | |
Running a 1060 and quite happy with it. If your laptop offers a Thunderbolt 3 connection, then any number of eGPUs are available. https://www.anandtech.com/show/15143/gigabyte-aorus-rtx-2080-ti-gaming-box-liquidcooled-tb3-egfx-graphics | |
ID: 53078 | Rating: 0 | rate: / Reply Quote | |
Thank you both. | |
ID: 53082 | Rating: 0 | rate: / Reply Quote | |
It does. Daisy chain allowed or is it too ambitious? | |
ID: 53083 | Rating: 0 | rate: / Reply Quote | |
From that egpu review, it seems that the gpu maxes out the TB3 connection. I doubt it could support another egpu daisy chained behind the first egpu. | |
ID: 53094 | Rating: 0 | rate: / Reply Quote | |
Has anyone used a "GTX 1650 SUPER" or "GTX 1660 SUPER" on GPUgrid? | |
ID: 53182 | Rating: 0 | rate: / Reply Quote | |
If you like, I can give you some more data about other "low end" cards for the NEW ACEMD3 application: | |
ID: 53188 | Rating: 0 | rate: / Reply Quote | |
Thanks for your feedback Carl | |
ID: 53198 | Rating: 0 | rate: / Reply Quote | |
Agree. Kudos Carl for defining what to expect from low end Nvidia cards. Even multi generation old cards are still usable for this project. | |
ID: 53202 | Rating: 0 | rate: / Reply Quote | |
Remember that the deadline is 5 days (120 hours) and the present workunits are quite short, as the fastest card (the RTX 2080 Ti at the moment) can finish them in 1h 42m. It can be 8-12 hours on the fastest cards by definition, so the long workunits can be 4.72~7.08 times longer than the present test workunits. If the workunits will get long enough to hit the 8h on the fastest cards, none of the cards on this list will be fast enough to finish them within the deadline. The deadline won't be extended. | |
ID: 53203 | Rating: 0 | rate: / Reply Quote | |
Looking back at the original post and subsequent comments...the conversation and quest for data and efficiency from GPUs has not changed in 10 years. | |
ID: 53204 | Rating: 0 | rate: / Reply Quote | |
I think the definitions of "long-run" tasks and "short-run" tasks have gone away with their applications. Now only New ACEMD3 tasks are available and in the future. | |
ID: 53206 | Rating: 0 | rate: / Reply Quote | |
I have some stats for the mining type cards: P102-100 and P106-100 CUDA: NVIDIA GPU 0: P106-100 (driver version 440.26, CUDA version 10.2, compute capability 6.1, 4096MB, 3974MB available, 4374 GFLOPS peak) CUDA: NVIDIA GPU 1: P102-100 (driver version 430.40, CUDA version 10.1, compute capability 6.1, 4096MB, 3964MB available, 10771 GFLOPS peak) http://www.gpugrid.net/show_host_detail.php?hostid=509037 There are 3 GPUs, units are minutes Dev# WU count Avg and Std of avg GPU0 WUs:6 -Stats- Avg:279.7(0.63) GPU1 WUs:6 -Stats- Avg:258.9(1.46) GPU2 WUs:6 -Stats- Avg:325.2(0.38) GPU0 is 1660Ti GPU1 is P102-100 GPU2 is 1070 http://www.gpugrid.net/show_host_detail.php?hostid=517762 There are 3 GPUs, units are minutes Dev# WU count Avg and Std of avg GPU0 WUs:2 -Stats- Avg:474.3(0.46) GPU1 WUs:2 -Stats- Avg:477.4(0.38) --this should have been the fastest card of the 3 GPU2 WUs:2 -Stats- Avg:497.3(0.59) GPU0 & 2 are P106-100 GPU1 is eVga 1060 with 6gb | |
ID: 53207 | Rating: 0 | rate: / Reply Quote | |
I think the definitions of "long-run" tasks and "short-run" tasks have gone away with their applications. Now only New ACEMD3 tasks are available and in the future. @TONI: would you please answer this above assumption. I have my RTX 2080 set for ACEMD3 only and my 2 GTX 1060s set for "Long" and "Short" WUs only. But my 1060s have not received a task in many days.Also why not update the GPUGrid preferences selection options to display reality.I realize this is not the best forum to address the situation but maybe it will be answered anyway. Billy Ewell 1931 (as of last week I can legally boast of being 88 great years of age!) Life is GOOD. | |
ID: 53208 | Rating: 0 | rate: / Reply Quote | |
Remember that the deadline is 5 days (120 hours) and the present workunits are quite short, as the fastest card (the RTX 2080 Ti at the moment) can finish them in 1h 42m. It can be 8-12 hours on the fastest cards by definition, so the long workunits can be 4.72~7.08 times longer than the present test workunits. If the workunits will get long enough to hit the 8h on the fastest cards, none of the cards on this list will be fast enough to finish them within the deadline. The deadline won't be extended. Thanks for the warning. :-) I think I wanted to post these in another post about older GPUs, but I need to take a look if I can find it again. I have a GTX1060 Founders Edition lying around here, but I still need to build a pc around it. :-) I'll look for a second hand desktop somewhere with a good enough power supply for the GTX1060. I hope that that card will be good enough for GPUGRID for at least the next 2-3 years. Cheers; Carl | |
ID: 53209 | Rating: 0 | rate: / Reply Quote | |
Agree. Kudos Carl for defining what to expect from low end Nvidia cards. Even multi generation old cards are still usable for this project. You are welcome. :-) I posted these values, most of all because it's always nice for newcomers to have figures which can give you an idea how long your own GPU will need to finish a work unit, especially if it is 'not so new' anymore. | |
ID: 53210 | Rating: 0 | rate: / Reply Quote | |
I must say that a deadline of just 5 days is the shortest I have seen in any distributed project. | |
ID: 53211 | Rating: 0 | rate: / Reply Quote | |
Depends on the project. Storage limits or the size of the database to hold tasks and work unit results have to be considered. | |
ID: 53212 | Rating: 0 | rate: / Reply Quote | |
Depends on the project. Storage limits or the size of the database to hold tasks and work unit results have to be considered. Thank you Keith! Thanks to information like this, I slowly learn how these distributed computing projects work in difficult circumstances: With a lack of funding, resources and manpower. Now that I think of it: These are exactly the cirumstances in which almost all scientists and doctors worldwide need to work. Because our governments prefer to invest in weapons rather than science and healthcare... So let us volunteers try to compensate. Warm greetings from a cold and wintery Brussels ;-) Carl Philip | |
ID: 53215 | Rating: 0 | rate: / Reply Quote | |
but why not extend the deadline to 6 days or even a week?The reason for this short deadline has been discussed many times. The work is generated from the previous result, so if the project extends the deadline by 2 days, the whole simulation could take up to (2 * the number of stages in the simulation) days longer. If you look closely you can distinguish the different parts of a workunit's name: test159-TONI_GSNTEST3-7-100-RND0962_0 The most obvious is the last digit: it's the number of re-sends.The "7-100" means it's the 7th stage of the 100 stages. Every stage is a workunit, the next stage picks up exactly where the previous finished. If the deadline would be 2 days longer for each stage, the whole series could take up to 200 days longer. The present 2999 workunits in the queue are actually 299.900 workunits. Different batches are split up to different stage sizes, depending of how urgent is the actual simulation for the project. If the present simulation have been split up to only 25 stages, the workunits have been 4 times longer. Less stages * short deadline = short total runtime More stages * long deadline = long total runtime | |
ID: 53217 | Rating: 0 | rate: / Reply Quote | |
I must say that a deadline of just 5 days is the shortest I have seen in any distributed project.At WCG Fighting AIDS at Home 2 is 24 hours and Africa Rainfall is likely to shorten to about 24 hours. I know it's frustrating but these molecular modelling projects require an enormous amount of work so they need to tighten it up so they can get the results and wet bench them. ____________ | |
ID: 53218 | Rating: 0 | rate: / Reply Quote | |
I think Rosetta has some now and then with a 3 day deadline. At least they have in the past. | |
ID: 53221 | Rating: 0 | rate: / Reply Quote | |
Could anybody please tell me whether it worth to buy a RTX 2080Ti GPU for Gpugrid, or may be any other modern but cheaper GPU will look more attractive in terms of сredit per €/$? | |
ID: 53438 | Rating: 0 | rate: / Reply Quote | |
Probably not. Find someone with a 2080Ti and look at the task times and then find someone with a 2080 Super and look at their task times. Then compare the percentage improvement of the 2080Ti to the percentage increase in cost. | |
ID: 53439 | Rating: 0 | rate: / Reply Quote | |
Message boards : Number crunching : Credit per € / $