Message boards : Graphics cards (GPUs) : What card?
Author | Message |
---|---|
Hi! | |
ID: 1888 | Rating: 0 | rate: / Reply Quote | |
Hi! a) Geforce 8400GS - but better run it on a single core PC, as otherwise the other units won't finish before the deadline, because the card is sooo slow. Only few shaders at a low clock speed b) Geforce GTX280 - if you can afford it | |
ID: 1889 | Rating: 0 | rate: / Reply Quote | |
Hi! Thanks. Ok. c) What is the best possible (not too slow, not too expensive) graphics card for crunching PS3grid.net workunits? Henri. ____________ | |
ID: 1890 | Rating: 0 | rate: / Reply Quote | |
Hi! 8800GT 512MB gdf | |
ID: 1894 | Rating: 0 | rate: / Reply Quote | |
Thanks! :) So, is this the correct model? Does this card support CUDA 2? Will it work fast enough in PCI Express 1.1? I only have PCI Express 1.1 motherboard. Where to get the latest drivers for that card? I have no experience on using NVIDIA cards (I have an ATI card at the moment). Henri. ____________ | |
ID: 1896 | Rating: 0 | rate: / Reply Quote | |
I recomended 8800 GTS 512 - 128 procesors.... It is 2.0 PCI-E but it should work without problems on PCI-E 1.0 drivers: http://www.nvidia.com/object/cuda_get.html | |
ID: 1901 | Rating: 0 | rate: / Reply Quote | |
I run 3 8800GT, 1 in each computer. They work very well and I see not much slowdown with normal windows operation. There is some noticed depending on what I'm running, most mostly not too bad. Some people have reported they can't even use their computer when it runs here. My two at work I use all day, heavily with little noticed slowdown and they are both P4-HT, running full boinc using both CPU and GPU. These were well priced for me, not to expensive like the high end cards. Mine are XFX brand and are single slot wide, an important thing to consider as some computers, especially all mine, can not take double wide cards without sacrificing another PCI slot, which I could not do as I have other boards and no empty slots to move them to. These are PCIe x16 2.0 but that is backward compatible with PCIe x16 1.1 slots. They worked fine in my PCIe x 16 1.1 slots. | |
ID: 1903 | Rating: 0 | rate: / Reply Quote | |
I think the FX9800GTX+ 512MB is a good (ot better) solution. You have to pay 10 EURO more but you will get more power. Isn't it? Or am I wrong? | |
ID: 1905 | Rating: 0 | rate: / Reply Quote | |
To be honest you can buy 8800GTS512 much cheaper then 9800GTX+ On both cart it is G92 the rally diference betwean this 2 cards is that 9800+ has smaller chip on 55nm and 8800 on 65nm and 9800+ is slighty faster due to aditional Mhz on core and memory. So price/performance better is 8800GTS, but on thermal 9800+ | |
ID: 1906 | Rating: 0 | rate: / Reply Quote | |
Thanks again! | |
ID: 1908 | Rating: 0 | rate: / Reply Quote | |
So, what exact make and model is quiet enough? There do exist 8800-series cards with passive cooling systems (ie no fan) but do expect to pay a premium for these! So long as the devices conform to Nvidia's reference designs, we'd expect them to be OK for GPUGRID. MJH | |
ID: 1909 | Rating: 0 | rate: / Reply Quote | |
Regarding noise I'll refer to the thread I just created. | |
ID: 1919 | Rating: 0 | rate: / Reply Quote | |
And let's get serious about the "most effective card" question. | |
ID: 1921 | Rating: 0 | rate: / Reply Quote | |
And let's get serious about the "most effective card" question. It is very intresting what you wrote. I have also a Q6600 overclocking at 3 GHz and will buy me a 9800GTX+ to morrow. The contigent of 1 WU per CPU and day seams me very small. In your calculation it should be 2 WUs | |
ID: 1924 | Rating: 0 | rate: / Reply Quote | |
Yes, actually I have a hard time to establish a 2 days cache.. but the GPU did not yet run dry :) | |
ID: 1925 | Rating: 0 | rate: / Reply Quote | |
Once upon a time... 2 weeks ago... we all happy owners of: | |
ID: 1926 | Rating: 0 | rate: / Reply Quote | |
...I'd like to know how fast a GTX260 is "in real world", because it's considerably more expensive than the 9800GTX+, the most expensive G92 card, but has about the same maximum GFlops. Well, I was able to run a few tasks on my GTX 260 with an earlier app version in the first tests under Linux64, but couldn't crunch more than one WU in a row because of driver problems, therefore I switched it to the Vista box... But as for the speed comparision - My EVGA GTX 260 was as fast as my EVGA 9800 GTX SC (super clocked), actually a little bit slower! ____________ pixelicious.at - my little photoblog | |
ID: 1927 | Rating: 0 | rate: / Reply Quote | |
But as for the speed comparision - My EVGA GTX 260 was as fast as my EVGA 9800 GTX SC (super clocked), actually a little bit slower! Thx! So the architectural fine tuning (more registers etc) of the GT200 doesn't yield any benefits (yet) for GPU-Grid and these cards have thus a rather bad performance per money. If I put the numbers in for the GTX280 I get 2000 credits/day more than a 9800GTX+ for 200€ more. Not a terrible deal, but I wouldn't recommend it. And I forgot the 9800GX2! 1TFlops for 260€ -> 8900 credits/day, 1600 credits/day more than the 9800GTX+ (assuming 1000 cr/day for both CPU cores) for 100€ more. Downside of this card is that it needs a 6 pin and an 8 pin power plug and aftermarket cooling solutions likely won't work due to the 2 chip architecture. MrS ____________ Scanning for our furry friends since Jan 2002 | |
ID: 1929 | Rating: 0 | rate: / Reply Quote | |
And let's get serious about the "most effective card" question. I just sold my 8800 GS and went for a 280, I dont regret the upgrade despite many complaining of the affordability! Real world benchmarks with folding@home and ps3grid and future mark showed me a 3x gain since the upgrade. I've sold my 8800 GS for 1/3 rd the price of the 280. "Flops" are misleading, I think the number of stream processors plays a bigger role, and frankly, I was never a big fan of SLIs. | |
ID: 1931 | Rating: 0 | rate: / Reply Quote | |
"Flops" are misleading, I think the number of stream processors plays a bigger role, and frankly, I was never a big fan of SLIs. Well.. no. Flops are calculated as "number of shaders" * "shader clock" * "instructions per clock per shader". The latter one could be 2 (one MADD) or 3 (one MADD + one MUL), but it's constant for all G80/90/GT200 chips. So Flop are a much better performance measure than "number of shaders", because they also take the frequency into account. And SLI.. yeah, just forget it for games. And for folding you'd have to disable it anyway. MrS ____________ Scanning for our furry friends since Jan 2002 | |
ID: 1932 | Rating: 0 | rate: / Reply Quote | |
Hello | |
ID: 1933 | Rating: 0 | rate: / Reply Quote | |
I've one 8800GT and one GTX280 running. The 8800GT needs 11:39h for one WU and I got 1987.41 credits, that's ca. 170 cr/h. The card works on a AMD Penom 9850 BE. The GTX280 needs only 7:50h for one WU and I got 3232.06 for it, that's ca. 415 cr/h. Are this other WUs, or why the credits are higher? | |
ID: 1937 | Rating: 0 | rate: / Reply Quote | |
I've one 8800GT and one GTX280 running. The 8800GT needs 11:39h for one WU and I got 1987.41 credits, that's ca. 170 cr/h. The card works on a AMD Penom 9850 BE. The GTX280 needs only 7:50h for one WU and I got 3232.06 for it, that's ca. 415 cr/h. Are this other WUs, or why the credits are higher?It's the new credit award with app v6.42, your 8800GT will also start to get 3232/WU when it runs v6.42 (or higher) ;-) | |
ID: 1939 | Rating: 0 | rate: / Reply Quote | |
Kokomiko, | |
ID: 1940 | Rating: 0 | rate: / Reply Quote | |
It's the new credit award with app v6.42, your 8800GT will also start to get 3232/WU when it runs v6.42 (or higher) ;-) Both machines runs with v6.43, what's wrong? ____________ | |
ID: 1941 | Rating: 0 | rate: / Reply Quote | |
Kokomiko, Both are not overclocked. The 8800GT (Gigabyte, 112 shader, 1728 MHz) runs under Vista 64bit on a Phenom 9850 BE with 2.5 GHz, the GTX280 (XFX, 240 shader, 1296 MHz) runs also under Vista 64bit on a Phenom 9950 with 2.6 GHz. ____________ | |
ID: 1943 | Rating: 0 | rate: / Reply Quote | |
The last WU finished by your 8800GT was still using 6.41, hence the lower credits. | |
ID: 1944 | Rating: 0 | rate: / Reply Quote | |
Which driver are you using? The newest, 177.84. ____________ | |
ID: 1945 | Rating: 0 | rate: / Reply Quote | |
Same for me. Now the only differences are that you use Vista 64 vs XP 32 for me and my Q6600 @ 3 GHz on a P35 board versus your Phenom. But this shouldn't have such strong effects. | |
ID: 1946 | Rating: 0 | rate: / Reply Quote | |
The last WU finished by your 8800GT was still using 6.41, hence the lower credits. It is whell known that some 3D prodecents make 3D cards with higher clocks that references... | |
ID: 1952 | Rating: 0 | rate: / Reply Quote | |
So the 8800GTS 512 seems to be the most efficient card of these. However, you have to take into account that you also need a PC to run the card in and one CPU core. I'll use my main rig to give an example of what I mean by that. I do not understand. Cannot I run two SETI@home WUs + gpugrid all at once with my dual core Pentium D 920 (and NVIDIA)? Henri. ____________ | |
ID: 1954 | Rating: 0 | rate: / Reply Quote | |
So the 8800GTS 512 seems to be the most efficient card of these. However, you have to take into account that you also need a PC to run the card in and one CPU core. I'll use my main rig to give an example of what I mean by that. No you can not. Eg. I have Q6600 and 8800GTS, and i crunch Rosetta@home and PS3Grid in TomaszPawelTeam :) So on 3 cores runs Roseta and on 1 core runs PS3Grid. At your computer on 1 core will be Seti@home and on second core will be PS3Grid .... I know to me it is strange too, and it shows that GPU is very powerful, but it need help from CPU to crunch.... So one core is always wasted to one GPU... P.S. If I have more $$$ i will buy 280GTX... but i dont't have so i crunch at 8800GTS512... If you have $$$ :) you should buy 280GTX :) | |
ID: 1955 | Rating: 0 | rate: / Reply Quote | |
It is whell known that some 3D prodecents make 3D cards with higher clocks that references... Sure. The point is that he's about 30 min faster than me with 112 shaders at 1.73 GHz, whereas I have 128 shaders at 1.83 GHz. That's a difference worth investigation. My prime candidate would be the Vista / Vista 64 driver. And, yes, currently you need one CPU core per GPU-WU. It's not doing any actual work, just keeping the GPU busy (sort of). MrS ____________ Scanning for our furry friends since Jan 2002 | |
ID: 1956 | Rating: 0 | rate: / Reply Quote | |
It is whell known that some 3D prodecents make 3D cards with higher clocks that references... My wife has on a Phenom 9850 BE a MSI 8800GT running under XP32bit, shader is running with 1674 MHz and she need 13:40h for one WU. Seems also to be faster then the stock frequency, but much slower then my card under Vista 64. ____________ | |
ID: 1957 | Rating: 0 | rate: / Reply Quote | |
GDF said going with the 2.0 CUDA compilers had a 20% performance hit under Win XP, which can be improved by future drivers. The Vista driver is different from the one for XP. So it seems the Vista driver got less than a 20% performance hit. | |
ID: 1959 | Rating: 0 | rate: / Reply Quote | |
"Flops" are misleading, I think the number of stream processors plays a bigger role, and frankly, I was never a big fan of SLIs. Remember the flops formula is the best the GPU can do (peak), but very few real world applications can issue the max instructions every cycle, unless you just have an application adding and multiplying useless numbers to maintain the maximum. | |
ID: 1960 | Rating: 0 | rate: / Reply Quote | |
Thanks for the info once again, guys! | |
ID: 1963 | Rating: 0 | rate: / Reply Quote | |
It is not wasted. If I understood it right the CPU is needed to feed the GPU with data... | |
ID: 1965 | Rating: 0 | rate: / Reply Quote | |
Remember the flops formula is the best the GPU can do (peak), but very few real world applications can issue the max instructions every cycle Yes, we're only calculating theoretical maximum Flops here, the real performance is going to be lower. This "lower" is basically the same factor for all G8x / G9x chips, but GT200 received a tweaked shader core and could therefore show higher real world GPU-Grid-performance with the same Flops rating. That's why I asked for the GTX260 :) Edit: and regarding CPU usage, F@H also needed 100% of one core in GPU1. The current GPU2 client seems tremendously improved. Maybe whatever F@H did could also help GPU-Grid? MrS ____________ Scanning for our furry friends since Jan 2002 | |
ID: 1967 | Rating: 0 | rate: / Reply Quote | |
Remember the flops formula is the best the GPU can do (peak), but very few real world applications can issue the max instructions every cycle I really hope improvements can be made in the future so more and more GPU computing will be available. I also hope more projects would try to build GPU applications so we are able to use full hardware potential for calculations. | |
ID: 1969 | Rating: 0 | rate: / Reply Quote | |
In every of my finished workunits I find the following text: | |
ID: 2036 | Rating: 0 | rate: / Reply Quote | |
What I´m wondering about is the MDIO Error Message. The restart.coor file that the error mentions is created when the processing of the work unit is suspended and stores the state of the simulation so that it may be restarted later. MJH | |
ID: 2037 | Rating: 0 | rate: / Reply Quote | |
I have the same card, but mine needs ~58600 seconds for one work unit. | |
ID: 2039 | Rating: 0 | rate: / Reply Quote | |
I totally agree with koschi. | |
ID: 2041 | Rating: 0 | rate: / Reply Quote | |
But it seems, that 21.5h computing time on a GF9800GT is a little bit to slow, isn´t it? Yes, I think so. I have 10,5 h with a GF9800GTX+ | |
ID: 2042 | Rating: 0 | rate: / Reply Quote | |
Back in the old days (hmm, 2 weeks ago?) I once managed to force the card to Powermizer level 3, but then the improvement was only from 48.000 to 47.000. Those were the days of the faster Cuda 1.x 6.25 app. | |
ID: 2043 | Rating: 0 | rate: / Reply Quote | |
Yes, his 9800GT is way too slow. Either the software is horribly broken / distrubed or it's not running at stock clocks (->GPU-Z). | |
ID: 2044 | Rating: 0 | rate: / Reply Quote | |
My 9800GT is running at stock speed: | |
ID: 2047 | Rating: 0 | rate: / Reply Quote | |
My 9800GT is running at stock speed: Update: With the Shaders overclocked @1750Mhz -- a WU took ~ 68000s (~19h) on my GF9800. Everything about that card seemd to be normal. GPU-Z detected 112shaders and the memorybus and clockspeed was OK, too. I updated the Video Driver to 177.92, but crunching times didnßt change. Yesterday I kicked the GT9800 and put in a GTX260. It´s a pity the 6.45APP was out on that day, so I couldn´t test a WU on 6.43 with it. With the 6.45 a WU takes ~10h -- that seems to be OK -- not as fast as some reported with the 6.43 (7-8h), but not such a big difference as my 9800GT did show. It seems there wasn´t anything wrong on my system, but the grafixcard was at fault. The 9800GT was a Gainward Bliss 9800GT (no golden sample -- just a normal standard card) Happy Cruning... Regards, Thorsten | |
ID: 2393 | Rating: 0 | rate: / Reply Quote | |
I also have a Gainward Bliss 9800GT 512MB, but mine is 20000 fas^ter than yours was. No idea why yours was so slow... | |
ID: 2395 | Rating: 0 | rate: / Reply Quote | |
Kyle, do you have another PC to put the 9800 in? If it's also slow there we can surely blame a "dodgy" card. | |
ID: 2400 | Rating: 0 | rate: / Reply Quote | |
No, I have taken them by their word on the 14day returning right and sent the 9800GT back .... :) | |
ID: 2406 | Rating: 0 | rate: / Reply Quote | |
- GTX280 has 933 GFlops for 350€ -> 2.66 GFlops/€ GTX280 350€ in the April of 2008? Can I soon after the Christmas of 2008 buy passively cooled GTX280 with just 300 €? Is it completely impossible to have passively cooled (and stable) GTX280 with 100% GPU load 24/7? Merry Christmas! HTH. ____________ | |
ID: 4452 | Rating: 0 | rate: / Reply Quote | |
Maybe GeForce GTX 285. It should run cooler, shouldn't it? :) Henri. ____________ | |
ID: 4487 | Rating: 0 | rate: / Reply Quote | |
GTX280 350€ in the April of 2008?[quote] If you find that fairy tell her I'll also take one of those.. but I'd like mine to be 100€. No seriously, Kokomiko was right when he said you can totally forget about cooling a GTX 280 passively (under load and in a computer case).. that's why noone else commented. You can easily handle 50W passively, but 70 - 80W gets painful, i.e. the card runs very hot and fails without sufficient case ventilation (which negates the benefits of running passive..). A GTX280 has 150+W power consumption.. do I need to say more? And, btw., 55 nm isn't going to help much. MrS ____________ Scanning for our furry friends since Jan 2002 | |
ID: 4503 | Rating: 0 | rate: / Reply Quote | |
A GTX280 has 150+W power consumption.. do I need to say more? And, btw., 55 nm isn't going to help much. Ok. Are those fans too noisy on those cards? I don't want to buy a card that screams so much I never ever want to have a 100% GPU load again... PS. Water cooling is out of business. I don't want to mess with water and electricity. Henri. ____________ | |
ID: 4504 | Rating: 0 | rate: / Reply Quote | |
Henri, | |
ID: 4505 | Rating: 0 | rate: / Reply Quote | |
The ASUS cards have fairly quiet fans in my experience. My 9600 GSO has one of the big "glaciator" fans, and the noise increase over the power supply fan is negligible at most (I really can't tell a difference to be honest). Ok. Thanks, Scott and everyone! Henri. ____________ | |
ID: 4509 | Rating: 0 | rate: / Reply Quote | |
Look here for further information about cooling. I suppose the stock GTX 260 cooler would still be to noisy for my ears, but any card with an Accelero S1 and 2 slow 120 mm fans is *cool*. | |
ID: 4580 | Rating: 0 | rate: / Reply Quote | |
Message boards : Graphics cards (GPUs) : What card?