Is StarCraft II Killing Graphics Cards? 422
An anonymous reader writes "One of the more curious trends emerging from last week's StarCraft II launch is people alleging that the game kills graphics cards.The between-mission scenes onboard Jim Raynor's ship aren't framerate capped. These are fairly static scenes, and don't take much work for the graphics card to display them. Because of this, the card renders the scene as quickly as possible, which then taxes your graphics card as it works to its full potential. As the pipelines within your graphics card work overtime, the card will heat up and if it can't cope with that heat it will crash."
Ridiculous. (Score:5, Insightful)
Re:Ridiculous. (Score:5, Informative)
i was thinking the same thing, many games arent FPS-capped anyway, and even in capped games, gamers will put the settings up so high that the game wont run at the capped framerate all the time
Graphic cards should be able to cope with it, although i do believe that it is possible to load a GPU in such a way that more transistors are active at the same time then the manufacturer thought would happen.
So unless there are reports of thousands of melted video cards, i call shens
The Fix (Score:5, Informative)
Re: (Score:3, Insightful)
Since I haven't seen anyone else post the fix, I will: Add the following lines to your "Documents\StarCraft II\variables.txt" file: frameratecapglue=30 frameratecap=60 You can add them to the beginning, end, or wherever. The game doesn't care.
Wouldn't ticking off the vsync option in the in-game settings be an easier way to fix the problem?
I'm having a hard time picturing any overlap between systems where unlimited framerate is a problem and vsync could drop your frame rate too low. I mean, usually it's the high end cards that have heat problems, not the ones where vsync might drop you to 1/2 or 1/4 refresh rate... If you have a high end card in a system that can't cool itself enough to use the card to it's full potential, who's problem is that
Re:Ridiculous. (Score:5, Insightful)
There is a parameter used for most high-dissipation ICs (such as CPUs and GPUs) - It's called "thermal design power".
This is the absolute maximum amount of heat the card can dissipate under any circumstances (not counting overclocking). The nature and definition of TDP means it should be physically impossible for ANY software to ever cause the card to exceed TDP.
If you have a system that can't handle the card running at TDP, that's faulty design of your system, not whatever caused it to hit TDP.
Re:Ridiculous. (Score:5, Informative)
There is a parameter used for most high-dissipation ICs (such as CPUs and GPUs) - It's called "thermal design power".
This is the absolute maximum amount of heat the card can dissipate under any circumstances (not counting overclocking). The nature and definition of TDP means it should be physically impossible for ANY software to ever cause the card to exceed TDP.
If you have a system that can't handle the card running at TDP, that's faulty design of your system, not whatever caused it to hit TDP.
Many video cards can exceed their TDP through certain sequences of instructions, and the drivers include code to prevent this from occurring. There's been issues in the past where this filter wasn't perfect, and cards were destroyed, typically when executing GPU stress tests.
Parent is insightful. (Score:5, Informative)
Exactly. Agree. That's the story here for anyone confused; hardware can be killed through software through no real fault of the user. See for instance Furmark [google.com] which ATI tries to throttle by checking for its name! No, you don't have to overclock, no, it's not because your cooling is subpar or because of dust or anything else, it's because HWVs don't want to spend the ten cents or whatever to take away the 'can run over peak for a few seconds' capability.
They're knowingly releasing hardware that can't survive 'full throttle', and it's bullshit.
PS. Here's a 8800GT fried during SC2 [hardforum.com].
Re:Parent is insightful. (Score:5, Insightful)
>it's not because your cooling is subpar
If your hardware can undergo a heat-related failure, then you have substandard cooling. That's pretty much the definition of substandard cooling.
Re:Parent is insightful. (Score:4, Insightful)
There's a difference between "substandard" and "insufficient."
A standard cooling solution can be insufficient under the right circumstances, such as a card that doesn't rate TDP correctly and allows the card to exceed its published TDP through software. The cooling isn't the problem; the manufacturer is the problem.
Re: (Score:2)
On top of that, it has been possible to put heat sensors on the chip and throttle the clock in case of overheating for several years now. IIRC Intel introduced this with the Pentium 4, and in some PCs with poorly cooled 3GHz+ P4 models this throttling actually kicked in. Annoying for the users, but at least their systems did not die.
Maybe AMD/ATI and NVIDIA should copy that feature? (apologies for dissing them if they have actually done so).
Re: (Score:3, Interesting)
Re:Ridiculous. (Score:5, Informative)
Not so much design. A few other new games have had this issue (notably Star trek online).
TDP assumes, wrongly, that your card is perfectly clean, and that the fan controls are always correct, which might be the case on a reference designed card, but might not quite be the case on factory overclocked boards or if there have been aftermarket tweaks to the driver to adjust the fan speed (which is usually a noise problem).
You're also assuming the fan is still perfectly mounted (which it might not have been in the first place), and that sort of thing. The PSU needs to be able to feed enough juice to the card, the case needs to be properly ventilated (and ideally cleaned), and god knows what other bits you've got in this board. Lots of boards have a northbridge fan that sits directly above the (first) gpu nowdays.
As a developer there's a bit you can, and should be doing to prevent this sort of thing. This sort of problem happens a couple of ways. One is the 'draw a simple scene as fast as possible' scenario in SC so cap the framerate at something like 120. The other is basically constantly feed the card as much data as possible (some beta builds of STO and early release had this problem), that was basically a problem of not being able to fit a whole area/level in memory, or not wanting to cause a load screen, so you're maxing out your bandwidth to push data to the card, while at the same time letting the player fly around and shoot stuff (and said stuff shoot back). One of the things here is to do a better job of controlling what's being sent to the card in the first place (BSP trees for example). That's a problem that the card will render a scene to look correct even if you treat it badly, so you can sort of plod through development like that, but you shouldn't assume that the uncleaned 3 year old system one of your customers has will be as pristine as your development machines.
When driving a car you can 'floor it' for a few seconds, but if you left it that way your engine would eventually overheat, if you've ever gotten stuck in the snow or on ice you'll know what I'm talking about. GPU's are similar. When your comp starts or when you do specific things with an application they can run with all of it's parts at full power, but only for a little while. If you do that for too long eventually it will burn out.
Re:Ridiculous. (Score:5, Insightful)
Not so much design. A few other new games have had this issue (notably Star trek online).
TDP assumes, wrongly, that your card is perfectly clean, and that the fan controls are always correct, which might be the case on a reference designed card, but might not quite be the case on factory overclocked boards or if there have been aftermarket tweaks to the driver to adjust the fan speed (which is usually a noise problem).
You're also assuming the fan is still perfectly mounted (which it might not have been in the first place), and that sort of thing. The PSU needs to be able to feed enough juice to the card, the case needs to be properly ventilated (and ideally cleaned), and god knows what other bits you've got in this board. Lots of boards have a northbridge fan that sits directly above the (first) gpu nowdays.
That is still a design issue. Proper engineering includes appropriate margins for error to deal with the real world including things getting dirty and not being put together perfectly.
Not everything is designed to operate at 100% duty cycle, but in those cases, the duty cycle is well documented and there are usually mechanisms in place to prevent actual damage if the limits are exceeded. Note how in the inevitable car analogy, there are warning lights and significant physical warning signs (such as steam pouring out of the engine compartment to let you know you have exceeded the engine's design capabilities. Unless you're stupid enough to ignore that and keep pushing, the engine suffers no actual damage.
Imagine the outrage people might feel if the engine's design limits could be exceeded just cruising down the highway and the first sign of it was that the engine just stops running and never starts again. There would most certainly be a class action suit alleging that the engine was defective by design.
Re:Ridiculous. (Score:5, Funny)
$ glxgears
5791 frames in 5.0 seconds = 1158.177 FPS
7120 frames in 5.0 seconds = 1423.968 FPS
6801 frames in 5.0 seconds = 1360.132 FPS
7110 frames in 5.0 seconds = 1421.871 FPS
Nope. No meltdown. Totally BS...
wow! what card?, and then I realized (Score:3, Informative)
I too, ran GLXGEARS to check my framerate, and was pulling 3500 FPS on a 6 month old good card, and was wondering - "HOLY fuct! -what card do you have that runs that fast?"
And then I remembered you could shrink the screen, and get higher FPS
(makes glxgers screen tiny)
20,900 FPS
21,500 FPS
meh...
Re:wow! what card?, and then I realized (Score:5, Informative)
glxgears isn't a benchmark. Its only point is to verify that "yep, 3D acceleration is working" on a very basic level. And even that isn't working that well now that it's possible to run it at high speeds in software.
Re: (Score:3, Funny)
That's the point, moron! StarCraft is supposed to overheat at drawing a lot of low-detail scenes. Glxgears is the _PERFECT_ example of cranking out a rediculous amount of frame at the most totaly _BASIC_ scenery.
Now given that a lot of this shit does happen with _COMPOSITING_ all the time (Windows 7 Aero with blurey windows) and the fact that it _DOES NOT_ happen with _OPEN MUTHERFUCKING COMPUTE LANGUAGE_ then it might be the god damned _FRAME BUFFER_ you _TOTAL FUCKTARD_.
Now get the _FUCK_ of my lawn...
Re:Ridiculous. (Score:5, Interesting)
Re:Ridiculous. (Score:5, Interesting)
No kidding. SC2 may end up being more intense if it happens to be just the right balance so that the ROPs, TMUs, and shaders all get to work to near capacity, but same shit: If your card crashes the problem is your setup, not the game. For a demo that'll kick the crap out of your card heat wise, try Furmark. It is designed such to run the chip to its absolute limits and thus have maximum power draw. If your system bombs it isn't the demo that is wrong, it is your computer. Maybe you don't have enough power, maybe your ventilation is bad, maybe your GPU has a defect in it. Whatever the case an intense load that causes a crash is revealing a problem, not causing it. Your system should handle any load given to it.
Re: (Score:2, Funny)
Re: (Score:2, Interesting)
Normally I would agree, the graphics card should be able to handle anything that is thrown at it, but there is something to this story. I have a radeon 4850 with one of these zalman coolers on it http://www.quietpcusa.com/images/vf1000-led.jpg and my case is big with lots of cooling. I have used this exact same configuration to play countless games over the last year, including MW2, BC2, etc. and never had a single crash. But now my system is crashing at the SC2 menus. My brother's machine is doing the
Acquired dust over the last year? (Score:2)
try vacuuming your card's heat sink,
and then see if it happens
I'm just saying
Re: (Score:3, Insightful)
Re:Ridiculous. (Score:5, Insightful)
If a graphics card cant survive a tight loop like that then it's designed by an idiot. My intel processor can survive that state.. and NO it's not like putting the car in neutral and mashing the throttle. you are comparing a device with a rotating mass getting sped up PAST the designed rpm rate+time point that actually would not cause a problem because of the rev limiter kicking in.
electronics, my uneducated friend, are different as there are NOT any moving parts in it.. This may surprise you.
I am tired of electronics nowdays designed for highest profit and not for quality. Engineers for chip makers are complete fucking morons if this is really happening.
Re: (Score:2)
i don't think it's so much the engineers as the beancounters doing "engineering" with an excel spreadsheet.
and since those beancounters are usually higher on the company hierarchy, the engineers either obey or get shafted.
to use a car analogy (sorry!), engineering by marketing resulted in the the ford edsel, engineering by beancounting resulted in ford pinto.
Re:Ridiculous. (Score:5, Insightful)
you forget that most chips people buy these days are the equivelant of "off the rack" clothing. they are manufacturing rejects that are sold as lower end cards. just like Xeon CPU's and the top of the line $600 graphics cards being the only ones that are a result of "perfect" manufacturing. everything else from i Core to your $150 graphics cards are manufacturing rejects with circuitry disabled. it's not like there is a production process for every single SKU of the 20 or so that ATI/Nvidia sell at any one time.
it all comes off one line, tested, binned and then circuitry is disabled depending on the results of the testing
Re:Ridiculous. (Score:5, Insightful)
While that fact is interesting, if I bought a chip that says it could do X, I still expect it to live up to X. It doesn't matter if X is a reject from manufacturing Y. If they were Y-rejects that still can't handle X, don't sell it as such.
Bananas (Score:3, Insightful)
I'm not sure why the parent post is tagged as insightful, because it's nonsense. Yes, some CPUs and GPUs are actually more powerful chips with defective components which were disabled, but the majority are not.
Nor does a fault in one part of the chip somehow make it less reliable than any other; faults are typically random due to imperfections in the die which affect only one small part of the silicon, and the rest of the chip will work without any problems.
The suggestion that that every CPU or GPU 'comes o
Re: (Score:3)
If there are enough failures to create a whole market, there's something really wrong.
Sometimes there aren't. A lot of the time, particularly as time wears on and the manufacturing of a particular chip gets better, you end up getting a better chip than you expected (marked/limited as expected, though) and can overclock it further. Desktop i7s are a pretty good example of this, with most of the late-model i7-920s and all of the i7-930s being able to go 500 MHz above spec without even breaking a sweat (i.e., on stock cooling). Pretty much every one can hit 4 GHz with even a basic after-market
Don't make car analogies if you don't understand (Score:5, Informative)
This is not putting your car in neutral and laying on the gas, it is a meaningless comparison. GPUs have no problem rendering excess frames, lots of excess frames, and simply not making any real use of them. This is no more a problem than having a CPU run a computationally intensive test that doesn't do anything. There is no difference from a heat or function standpoint between all the units being fully active rendering something simple quickly or all the units being active rendering something complex slowly. In either case all the logic is active with lots of power flowing through and thermal output is maxed. A component should be able to handle this, no problem. Whatever a CPU or GPU is rated to for speed is not a temporary max, it is what it can run at full time. If there is a failure, it indicates a defect of some kind somewhere.
The most usual defect is inadequate airflow. People have a case with poor airflow, and reduce it further by not clearing dust buildup. As such the components can't cool themselves well enough.
As the GP said: This is a non-issue. If it happens to you, the game revealed a problem, it didn't cause it. Fix your system.
Re: (Score:2)
well to be honest, there are so few cases with good airflow, and then the GPU manufacturers skimp on the cooling power of the stock heatsinks. 12volts and 10 amps is 120 watts, so yes you will need some substantial cooling to keep things in spec temps. Although cases like the SG07 have dedicated cold air for the GPU and a fair amount of cooling on the backside of the card as well.
In the end though, this really should be a non issue, with proper component selection, placement and cooling.
Re: (Score:3, Funny)
The only way that TFA's "rendering simple scenes fast" could possibly be true is if modern graphics cards have some component that does something brief but power-hungry, once a frame, that is adequately heat-sunk for normal operation but which causes overheating at high frame rates. That said, TFA seems more to be pointing the finger at old, poorly assembled machines which have th
Re: (Score:3, Interesting)
How is it not the responsibility of the card vendor to engineer their cards so they won't overheat? To me this is black and white; if a laptop or video card melts when running a program that taxes some part of the system, unless you've gone out of your way to turn off sensors or block airflow, it's an engineering fault.
Re: (Score:3, Informative)
Not sure I get the reasoning here (Score:3, Interesting)
Are most games framerate-capped? Wouldn't all games, at all times, be rendering as quickly as possible, operating to the graphics card's full potential?
Re:Not sure I get the reasoning here (Score:5, Insightful)
To clarify my stance: This story is retarded, and all the time you look at it/think about it is time you won't get back.
Re:Not sure I get the reasoning here (Score:5, Insightful)
At this point I suspect "Kdawson" is a lot like "Alan Smithee". He just forgot to tick the box this time.
Re: (Score:2)
Re: (Score:2)
I still wonder if there is something more to this story than a bunch of cards with insufficient cooling crashing. Few if any professionally assembled PCs should have bad enough cooling that a game could cause the GPU to overheat.
Re: (Score:3, Interesting)
>>This story is just FUD and troll. I would've expected it to come from kdawson, but apparently I gave Taco too much credit.
It's news, because... it's about Starcraft 2? Kinda?
Why not run a story about how Quake 1 is killing modern computers? The last time I ran Quake it was somewhere above 300fps with vsync disabled.
Re: (Score:3, Insightful)
The problem with enabling vsync is the following:
A standard LCD has 60Hz, which is about 1.67ms per frame. When your game requires 1.7ms to render a single frame, without vsync that's about 58.8fps, which isn't that bad (you wouldn't notice it).
When vsync is enabled, what happens is that the first frame isn't ready when the screen is refreshed, so the whole pipeline stalls until the next vertical sync. On the next one, you can finally display your image, and render again for 1.7ms. The whole cycle repeats.
Re: (Score:2)
I find it odd that people don't want the game frame rate capped.
Why go past 60 FPS? Okay maybe 120 if you nuts?
What do you gain? I would rather not put out the heat and eat up the power.
Of course the only game I really play is FSX and I would love to see 60 FPS with all the eye candy turned on.
read: StarCraft will expose your crappy setup (Score:5, Informative)
Clearly StarCraft is not at fault here. No software should be capable of damaging your graphics card. But if the thermal design of your system is broken, then it's your fault, or the manufacturer's.
If your card breaks and there is nothing wrong with your cooling, then your card was already broken before you even fired up StarCraft.
Re:read: StarCraft will expose your crappy setup (Score:5, Insightful)
Or a more developed version of the same argument:
Starcraft 2 has a pretty wide audience, by the standards of a PC/Mac game, and while it's certainly not a Crysis-style hardware-hog, it does have higher requirements than a lot of the usual mass-market PC games (eg. The Sims and its sequels). In addition, its prequel, which is 12 years old and was technically underwhelming by the standards of its own time (the graphically-far-superior Total Annihilation actually came out first) has a large hardcore fanbase, a lot of whom probably don't play much other than Starcraft.
So Starcraft 2 is released and is promptly installed on a lot of PCs that are not routinely used for gaming, or at least for playing games less than a decade old. A large chunk of these PCs have never run a high-end modern game before. When asked to do so, the less-than-stellar graphics cards in a good portion of them give up and fall over. No conspiracy, no fault in Starcraft 2, just a lot of crusty PCs being taken outside of their comfort zone and not faring so well.
Uhh... (Score:5, Interesting)
You can uncap the framerate in lots of games, but we've never heard about this problem before. I don't think this is a problem. Especaily since you can easily make a GFX card run at full capacity and a low framerate by simply playing a game that's a little too new for it, something a lot of people trying to put off upgrades do. If your GFX card can't run at it's maximum capacity without overheating, something is wrong with its cooling.
Correction (Score:2)
s/I don't think this is a problem./This sounds like a red herring to me.
And the article talks about dust being the problem; which is exactly what I was thinking of when I said "something is wrong with its cooling". I've had that problem before with my old GPU; in Left 4 Dead 2 (but not in older games like TF2) I'd get great slowdowns every so often and my GPU was running pretty hot. Turns out it was throttling itself to keep itself from getting even hotter. A heatsink cleanout fixed that right up.
Re: (Score:3, Insightful)
Re: (Score:3, Informative)
Re: (Score:2)
Yeah I ripped apart the summary before I moved on to the article. It does look like the article still tries to place part of the blame on Blizzard though, as the author expects it to be patched.
it's pathetic for computer hardware to kill itself by overheating, but if you know that it can happen, you should still do your best not to overheat it.
Re: (Score:2)
Yeah, don't play games like Starcraft till you buy a proper graphics card that can handle it
If the cards were dying even when the fans were working fine,there wasn't that much dust, and the ambient temps were within range, then I'd say the cards are faulty.
Might explain my crashes (Score:5, Informative)
I'm playing Starcraft II on the last-gen iMac (purchased about four months ago) on OS X 10.6.3. The game is stable during gameplay, but it's crashed on me several times in cutscenes, onboard the Hyperion, or even in the main menu (ironically, while I was bringing up the menu to quit the game).
Re: (Score:2)
Re: (Score:3, Interesting)
Ever feel how hot those things get now, even when running normally? It's not surprising that they would completely fall over if you push them even moderately hard.
*Cough* (Score:2, Insightful)
Bullshit.
Software cannot kill Hardware (Score:2, Insightful)
Only lazy firmware developers for hardware can do that, the fault is not any game, its the driver (or, if the program somehow turns of the fan)
Already dead (Score:3, Insightful)
Graphics card that can't handle working to its full potential is already dead (as designed).
game was crashing for me (Score:5, Informative)
This may have been the problem I experienced. I had played in the (multiplayer only) beta with no problems. Once the game came out though, I kept crashing in single player in between levels. I cleaned the dust out of my computer and that solved the problem.
I wonder how many people experiencing this just have too much dust built up in their computers?
Re: (Score:2, Informative)
Seriously..? (Score:5, Insightful)
So it dies when used as intended? (Score:2, Insightful)
What year is this? (Score:5, Interesting)
When graphics cards overheat, the worst thing that happens is a blue screen. On ATI cards, they just restart the card (it does a recovery-mode type of thing).
You can overclock any card to insane temperatures (90C+) without them even turning off, much less breaking them. There is simply no way that Starcraft 2 is killing any graphics cards.
There *was* one issue with an nvidia patch a while back which a driver update actually did kill some graphics cards, but it was nvidia's fault, and they promptly fixed it.
This article is pure misinformation.
Re: (Score:2)
My old nVidia card would underclock itself when it started to overheat. Good thing too, when my heatsink got clogged with dust it probably saved itself when I was still clueless about it.
Even if its true... (Score:5, Insightful)
more than crash... damage (Score:4, Interesting)
The summary says an overheated video card will crash. It will do more than crash. It can permanently damage the video hardware. This seems like a major hassle to swap out the video components on a big gaming rig, but it can be a lot worse for high-end laptops. I've had similar problems with 3D software running on a MacBook Pro -- plenty of performance, but the video card gets second priority in the heat-management.
In my MBP, there are separate temperature probes on the CPU, hard drive, battery and chipset, but none on the dual video chip units, so the thermostat-controlled fan won't even kick in when either the "integrated" nor the "high performance" video units are the only stressed component.
Besides the hardware cooling problems, there's no reason for trying to draw more than 120 fps on most LCDs; software needs to get more responsible about heat and speed resource usage when given access to over-spec hardware. Limit the rendering loop to 90~120 fps, unless you're doing something purposely exotic such as driving stereoscopic displays or CAVEs (at 90~120 fps per camera).
Re:more than crash... damage (Score:5, Insightful)
I'm going to have to disagree here. It's not up to software developers to go around making sure hardware x and y won't just roll over and die during certain sections of their game.
It's up to hardware manufacturers to make sure their hardware works under all normal conditions. I mean really, if you make hardware that can fry itself, maybe you're pushing it to far.
Gee whiz guys! We can render this game at 4839483 FPS! But don't do it for more than 2 seconds or it'll melt! Woot, time to release them en masse! The benchmarks will look awesome!
Pushing a card to its max should _never_ cause it to "crash", let alone get damaged.
Re: (Score:3, Interesting)
Re: (Score:2)
In my MBP, there are separate temperature probes on the CPU, hard drive, battery and chipset, but none on the dual video chip units, so the thermostat-controlled fan won't even kick in when either the "integrated" nor the "high performance" video units are the only stressed component.
Sounds like a hell of a design problem. Given what you had to have paid for it, I'd take it back. There's no excuse for that kind of incompetence.
Either that, or you just don't know what the hell you're talking about. I give it 50/50 odds.
No. (Score:2)
No. Crappy cards that overheat when left running displaying ANYTHING (static images, top-end 3D, what does it matter?) are killing those graphics cards. CPU's (and therefore GPU's) should detect overheat, then throttle back or switch off as necessary. If that still causes a problem in 2010, you have bigger problems on your hands than how often a game decides to blit surfaces about - such as a potential fire. If your case is that dirty, your card should still cope anyway, even if that means it just overh
Short answer: No (Score:5, Funny)
..or... Are graphics cards poorly designed? (Score:2)
We're talking about a piece of hardware here which is capable of melting itself down with no internal cap on processing, and we're blaming the software?
IANAE (I am not an engineer) but it seems to me that the software designers should be able to throw whatever they like at the cards, and it's up to the hardware manufacturers to see to it that the hardware doesn't self destruct.
Orrin Hatch (Score:2)
Is this Orrin Hatch's "Destroy the PCs" [bbc.co.uk] plan made manifest? It has taken 7 years, but what subtle, indeed Machiavellian implementation.
Issue was fixed in patch 14 of Beta (Score:2, Interesting)
God /. you are WAY behind here. This was an issue 5 months ago in the Beta. There IS a hard cap in menus now.
BS (Score:2)
What about OpenCL/CUDA? These frameworks use the card's full potential, so far nobody reported any issues. If the card has cooling problems, it's clearly the faulty hardware. The only downside is a slightly more heat and noise from the videocard than there should be during these scenes. This is not a car where revving the engine on neutral indeed stresses the engine.
Mindless, driveling sensationalism (Score:4, Insightful)
OMG NEW HIGHLY ANTICIPATED TITLE KILLZ0RZ YOUR COMPUTAR!!!
No, if your machine is crappy, this exposes that you've got cooling or power problems, or both. You should see that you fix these.
In '94 I had a 486SX-25 that would choke and die when playing Doom in multi-player from time to time. It wasn't that the game KILLZ0RED MY COMPUTAR, it was that the CPU couldn't keep up with everything. Sticking a DX2-50 Overdrive into the socket solved that problem.
My guess? Users need to STFU (Score:5, Insightful)
I fail to see how rendering a scene at a high framerate would be any more challenging than rendering a complex scene at a lower frame rate. Remember that the hardware either is or is not in use. The ROPs, the shaders, etc. It isn't like there is some magic thing about a simple scene that makes a card work extra hard or something.
So my bet is you have users that have one or more things happening:
1) They are overclocking their cards. This is always a potential for problems. When you push something past its spec, you may find it has problem in some cases.
2) Their airflow sucks. They have inadequate ventilation in their case for their card.
3) Their PSU is inadequate for their card. High end graphics cards need a lot of voltage on the 12v rail. If you have one that can't handle it, well then maybe you have some problems in intense games.
Really, this sounds no different than the people who OC their processor and claim it is "perfectly stable" but then claim that Prime95 or LinX "break it." No, that means it is NOT perfectly stable, that means you have a problem. Doesn't mean the problem manifests with everything, but it means that you do have a problem that'll show up sometimes.
I'm betting it is the same thing here. It isn't that SC2 is "killing" their card, it is that their card has problem and SC2 is one of the things that can reveal that. There are probably others too.
So if your system is crashing in SC2 disable any overclocking, make sure you've got good ventilation (which may mean a new case) and make sure you have a PSU that supports your graphics card, including providing dedicate PCIe power connectors sufficient for it. Don't blame the software for revealing a flaw in your system.
Re: (Score:3, Interesting)
I fail to see how rendering a scene at a high framerate would be any more challenging than rendering a complex scene at a lower frame rate. Remember that the hardware either is or is not in use. The ROPs, the shaders, etc. It isn't like there is some magic thing about a simple scene that makes a card work extra hard or something.
Games now a days are highly threaded, with game logic and rendering happening in parallel and both in lock step (waiting for each other to finish). The difference between a complex scene and a simple scene is that the render thread will have less to update, and do more draw calls. If there is little or no animation to update (either updated in the game logic and pushed acr
Re: (Score:2)
Open case, put desktop fan roughly inside, full blast. If it stops crashing, get better ventilation in your case.
Re:My guess? Users need to STFU (Score:5, Funny)
Their PSU is inadequate for their card. High end graphics cards need a lot of voltage on the 12v rail.
I'd say they need 12V of voltage on the 12V rail...
Re: (Score:3, Informative)
I'm betting it is the same thing here. It isn't that SC2 is "killing" their card, it is that their card has problem and SC2 is one of the things that can reveal that. There are probably others too.
So if your system is crashing in SC2 disable any overclocking, make sure you've got good ventilation (which may mean a new case) and make sure you have a PSU that supports your graphics card, including providing dedicate PCIe power connectors sufficient for it. Don't blame the software for revealing a flaw in your system.
I guess we can all be glad you don't work for Blizzard. Here's what the pro's said:
Screens that are light on detail may make your system overheat if cooling is overall insufficient. This is because the game has nothing to do so it is primarily just working on drawing the screen very quickly. A temporary workaround is to go to your Documents\StarCraft II Beta\variables.txt file and add these lines:
frameratecapglue=30
frameratecap=60
You may replace these numbers if you want to.
Note how this is kind of the same thing, but Blizzard's solution has some actual tact behind it...
Re: (Score:3, Informative)
The difference is that with complex scenes, the framerate is limited either by the CPU (to calculate AI and physics) or IO (to send commands, textures and meshes to the card). Wit
The cards were already broken (Score:3, Insightful)
If a process, like a webserver, could erase itself from a hard drive by benign input, it would be a bug. This is no different.
My graphics card, a GTX 275, was factory locked to a 40% duty cycle on the fan, no matter how hot it got. I had to resort to RivaTuner to make the fan auto-adjust speed based on temperature. Since there is no speed limit where I'm putting people's lives at risk for rendering too many frames per second, or any other reasonable reason to limit the amount of work a card can do before it destroys itself when the hardware is perfectly capable of doing that work without destroying itself, the only conclusion is that it is defective.
That said, anything that doesn't use vsync is stupid, period, always, (unless you're benchmarking or trying to warm a cold room). Spending that extra processing power on a proper motion blur would have a far greater effect on perceived smoothness.
Re: (Score:2)
I agree that the card you mention had an issue. But the main problem is that die temperature sensing is such a simple thing to so. Power chips (switchers, regulators, bridges) that sometimes sell for $0.10 apiece can have die temperature sensors and can turn themselves off to cool down. Why the heck graphics chip makers don't put temperature-controlled power management (clock scaling, unit cycling) is beyond me. It's not like it's rocket science. If you know your engineering, you should even be able to have
Wait, why is this not consider a hardware flaw? (Score:2)
What year is this? (Score:2)
Re: (Score:2)
performance evaluation tools don't kill graphics.. (Score:2)
but saying that using your graphics card at full juice is essentially saying that pcmark& etc programs would do it as well. there's an issue with the card already if it's killing them. the real problem with the game is that it hasn't evolved at all from
JUST IN: Using your computer can reduce it's life (Score:2)
This is because StarCraft II is correctly written. (Score:2, Insightful)
StarCraft II is exposing shoddy thermal engineering in video cards because, unlike most games on the market, StarCraft II is correctly utilizing your video card to it's fullest potential.
Say what you will about SC2 game balance, say what you will about Battle.NET 2.0's crappy interface, say what you will about how cheesy Jim Raynor is. I wouldn't disagree with you.
But when it comes to writing engines, Blizzard is the best of the best. Hands down. Everything they write runs smooth as silk, and they have a ge
Re: (Score:2)
I agree. Thermal engineering bites even big names in the ass. Repeatedly. MacBook Air, anyone?
Re:This is because StarCraft II is correctly writt (Score:4, Informative)
Uh, no, eating as much GPU power as possible to render a static scene hundreds of times a second on a display that can only probably display 60 frames per second is not an example of properly-written software. In fact, it's just plain stupid, and nearly as wrong as you can possibly be.
That said, it shouldn't have any effect on graphics cards other than making less resources available to other concurrently-running programs, and Blizzard should in no way be blamed for breaking people's cards.
Evil Giant Killer Dust Bunnies From Hell (Score:4, Informative)
The summary should say that it's the Evil Giant Killer Dust Bunnies From Hell, not Starcraft, that are shutting down the cards.
The breaking test (Score:2)
Some building do this: at weekends, turn off and on the power of the whole building. This serve the purpose to force these bulbs that are about to fail.. to fail sooner. Maybe SC2 has one of the most popular games released in years is working as a unintended "break test". But it will be good if Blizzard adds some caps *anyway*.
Weird (Score:2)
Air Duster (Score:2)
College, Careers, Marriages (Score:3, Funny)
Mobile & "greener" gaming (Score:2)
I sincerely doubt that the summary's claim is true, any decent graphics card should either underclock itself to lower the heat generation, or simply BSOD on you. Furthermore, if scenes are very easily rendered, wouldn't this also automatically mean that less of the chip is used ? I'm no expert in graphics hardware design, but one would think (I know, bad idea) that they would have special parts of the silicon dedicated to some more complicated tasks, which in a case like this wouldn't be used at all. If tha
This is this way the entire game (Score:4, Insightful)
The entire game is not capped. It's been that way since beta started. The framerate cap variables have also been published from shortly after the beta came out.
Why Blizzard doesn't cap their games at 60fps (or hell 120fps if they think 60 is too low for some reason) I don't know. There's really no reason to render frames faster than that, even if you can.
Re: (Score:2)
Re: (Score:2)
Some game engines tie the physics calculations to the framerate, so having higher FPS in those does make the movement more realistic.
Additionally, most engines poll the input devices only once per frame. If you have a high-end mouse or gamepad that actually responds more than 60 times a second, having higher framerates will make input more fluid.
Finally, I like having a buffer zone. I usually tune my games to render about 80 fps, so if I run into a scene a bit more complex than normal, I still see 60 frames
Re: (Score:2)
Because graphics cards are used for many other things besides driving monitors. Think off-screen rendering (done by any modern OS behind your back), scientific calculations, etc.