Slashdot is powered by your submissions, so send in your scoop

 



Forgot your password?
typodupeerror
×
Graphics Upgrades Games Entertainment Hardware

Ubisoft Claims CPU Specs a Limiting Factor In Assassin's Creed Unity On Consoles 338

MojoKid (1002251) writes A new interview with Assassin's Creed Unity senior producer Vincent Pontbriand has some gamers seeing red and others crying "told you so," after the developer revealed that the game's 900p framerate and 30 fps target on consoles is a result of weak CPU performance rather than GPU compute. "Technically we're CPU-bound," Pontbriand said. "The GPUs are really powerful, obviously the graphics look pretty good, but it's the CPU that has to process the AI, the number of NPCs we have on screen, all these systems running in parallel. We were quickly bottlenecked by that and it was a bit frustrating, because we thought that this was going to be a tenfold improvement over everything AI-wise..." This has been read by many as a rather damning referendum on the capabilities of AMD's APU that's under the hood of Sony's and Microsoft's new consoles. To some extent, that's justified; the Jaguar CPU inside both the Sony PS4 and Xbox One is a modest chip with a relatively low clock speed. Both consoles may offer eight CPU threads on paper, but games can't access all that headroom. One thread is reserved for the OS and a few more cores will be used for processing the 3D pipeline. Between the two, Ubisoft may have only had 4-5 cores for AI and other calculations — scarcely more than last gen, and the Xbox 360 and PS3 CPUs were clocked much faster than the 1.6 / 1.73GHz frequencies of their replacements.
This discussion has been archived. No new comments can be posted.

Ubisoft Claims CPU Specs a Limiting Factor In Assassin's Creed Unity On Consoles

Comments Filter:
  • Linked? (Score:4, Interesting)

    by Fwipp ( 1473271 ) on Thursday October 09, 2014 @03:05PM (#48105971)

    "We could be running at 100fps if it was just graphics, but because of AI, we're still limited to 30 frames per second."

    Uh, have you guys tried running the AI calculations less frequently than graphics redraws? You don't have to keep them in sync, you know.

    • I'm sure they have thought of that
      • Re:Linked? (Score:5, Insightful)

        by lgw ( 121541 ) on Thursday October 09, 2014 @03:23PM (#48106183) Journal

        Some engines don't give you much choice. I'd hope modern games aren't still stuck in the single-threaded, hard-clocked world of yesteryear, but you never know.

        It's also possible they're using some very slow high-level language for the AI, and/or that no one's ever done an algorithmic optimization pass on the AI code, and they just couldn't keep up with the pipeline of collision events and whatnot that are often tightly coupled with framerate.

        I've been amazed in some MMOs how server performance will be totally trashed by some patch to the AI, and completely restored by the next. Poorly thought out AI code can certainly bring a CPU to its knees.

        • > I'd hope modern games aren't still stuck in the single-threaded, hard-clocked world of yesteryear, but you never know.

          Nah, game devs were forced to multithread their engine back in the PS3 and Xbox360 days. (In the PS2 days you had to offload data to the VU, GS, IOP, and SPU.)

          Current gen consoles are basically a dumbed down x86 that have 8 threads.

          • Number of threads less important than how you manage the cache. And you have to manage the cache, avoiding false sharing [wikipedia.org], in order to get near 6-7 x performance with 8 threads than you would have got with 1. If you don't take care of your cache, you'll run 8 threads and it'll be 1.5 x faster and you'll wonder why. I suspect something like this is going on.

            It's very difficult with complex algorithms or structures of course. Relatively easy with parallel primitives if your tasks are simple.
        • It's also possible they're using some very slow high-level language for the AI, and/or that no one's ever done an algorithmic optimization pass on the AI code, and they just couldn't keep up with the pipeline of collision events and whatnot that are often tightly coupled with framerate.

          It's possible, but highly unlikely. Game AI doesn't have all that much in common with research AI, and the issue of optimisation has meant that game AI is generally coded very close to the metal (C++, quite often).

        • Re: (Score:3, Interesting)

          by Anonymous Coward

          Oh but every engine is still doing this.

          Crytek 3 ? Still locked to one CPU core
          Unity? Still locked to one CPU core

          Even though they may do "multithreaded rendering" this really only means the CPU part of the rendering splits passes. The entire calculations are still being done as though they were single threaded.
          In a single threaded game, I might do A,B,C,D,E in that order, but if I have more cores, I'll just stick all these on separate threads... and hope the damn OS schedules it properly. So instead of get

      • Re:Linked? (Score:5, Interesting)

        by donscarletti ( 569232 ) on Thursday October 09, 2014 @09:00PM (#48108357)

        I'm an engine programmer who has been lead on 2 published titles (PC not console).

        I'm pretty sure they thought of that too and already did it. One has to write multi-threaded code for these consoles since they are multi-cored and otherwise most of the resources are wasted. Multi-threaded code is really easy to arbitrarily set tick frequencies and lock contention on the rendering thread is actually lower when you set the tick frequency of things like physics and AI to a lower frequency to your FPS, especially if your FPS is not an even multiple of this frequency. We run Havok at 50hz and render at 60fps, it sounds counter-intuitive, but it looks and feels great.

        The things is though, this game is obviously either GPU limited or close to becoming GPU limited. The key here is not the 30fps, which without looking at profiling results, could be equally easily explained by CPU limiting or GPU limiting, but the resolution of 900p, which the CPU should have absolutely nothing to do with. So you cannot confidently say that it is GPU limited now, but it certainly would be at 1080p, otherwise they would have just upped the res without it slowing the framerate.

        The issue here is Vincent Pontbriand is probably not a technical guy. Roles between companies vary and it's hard to know who if anyone reports to a "Senior Producer". If the engine programmers reported to him, it would be possible that he was lied to, since explaining exactly why framerates are the way things are is often tiresome and complex, since bottlenecks can be in many different places in the GPU pipeline (geometry, shader, input, texture, ROP, framebuffer) and the position of the bottleneck may shift while rendering a single frame, the bottleneck can also shift between the CPU and the GPU during a singe frame if the instruction buffer fills up. As it is, they probably don't report to him, so I would say that he probably just doesn't know the whole picture.

    • Re:Linked? (Score:4, Interesting)

      by GameMaster ( 148118 ) on Thursday October 09, 2014 @03:13PM (#48106061)

      It's quite possible that he means they have artificially slowed down the graphics rendering to provide more cycles to the AI.

      • It's quite possible that he means they have artificially slowed down the graphics rendering to provide more cycles to the AI.

        This is how I read it as well. Though, pure rendering and lerping should not eat up much CPU especially on consoles. Unless they've got a really inefficient rendering pipeline. I'm curious exactly how much extra AI this would allow them to run.

    • Re:Linked? (Score:5, Insightful)

      by Forever Wondering ( 2506940 ) on Thursday October 09, 2014 @05:05PM (#48107027)

      It's probably not the AI calculations related to gameplay, but Ubisoft's AI calculations related to their DRM that get highest priority in their games ...

  • by Anonymous Coward on Thursday October 09, 2014 @03:06PM (#48105979)

    can we please stop pretending clockspeed has anything to do with performance?

    as has been said a million times

    CLOCKSPEED IS ONLY FOR MEASURING APPLES WITH APPLES.

    different CPU generations are INCOMPARABLE using clock speed.

    so for instance a sandy bridge at 2.0 ghz is slower than a haswell at 2.0ghz, even with the same ghz number.

    Benchmarking is the only way.

  • by schlachter ( 862210 ) on Thursday October 09, 2014 @03:12PM (#48106055)

    Ubisoft may have only had 4-5 cores for AI and other calculations — scarcely more than last gen, and the Xbox 360 and PS3 CPUs were clocked much faster than the 1.6 / 1.73GHz frequencies of their replacements.

    Clock speed is not a good comparison. These processors should process data much faster than those in the Xbox 360 and PS3 despite the clock speed differences.

    CPU that has to process the AI, the number of NPCs we have on screen

    Maybe you didn't target your game properly.

    Many Nintendo games run in 1080p and 60 frames/second on the Wii U which is much less powerful...because Nintendo makes that their target when deciding how much AI and graphics detail to put on the screen at once.

    • Which is why I usually prefer playing games on Nintendo systems. When I play with my friends on their Xbox 360, Xbox One or PS4, the games may look good but there's often a really huge drop in the frame rate, I don't know how they can enjoy games that way.

    • by AmiMoJo ( 196126 ) * on Thursday October 09, 2014 @03:55PM (#48106507) Homepage Journal

      Maybe you didn't target your game properly.

      I think what Ubisoft is trying to say here is that it's programmers are shit, and so is it's game engine (AnvilNext). Other publishers manage to do okay, but poor old Ubisoft are stuck with this turd and can't just switch to Unreal or something more competent, so sorry guys you only get 900p on consoles.

      I mean, obviously if you are CPU bound the solution is to reduce the load on the GPU by making the game render at 900p. The problem is the AI for the large number of characters in the game, so clearly reducing the pixel count will help with that.

      • by tibit ( 1762298 )

        What if the consoles have shared system memory and they're hitting the memory bandwidth with the graphics?

      • They also might be running into an skinning bottleneck. Are they using 4 bones / vertex? How many AIs do they have onscreen at once? What is their poly count per character?

        One of their old games, RB6:LV2 doesn't even draw more then 8 characters maximum at any one time, even though the level clearly supports 50+ terrorists.

      • by AmiMoJo ( 196126 ) *

        Gah... should be "its game engine". Forced to troll myself now.

    • by ravyne ( 858869 )
      Clockspeed really isn't a good comparison. The PPC chips in the last-gen consoles were the PowerPC equivilent of Intel's first-generation Atom processors. They were more narrowly-issued than Jaguar, and didn't have fancy branch predictors or instruction re-ordering; like Atom, they used hyperthreading to hide when one thread stalled, but that only goes so far. Jaguar in the current gen has a nice branch predictor, modest instruction re-ordering, can issue more instructions per cycle, and also supports AVX.
      • by tibit ( 1762298 )

        "right about unified memory architectures" So, that pretty much settles it for me: their shaders push the memory subsystem to its knees and at 1080p the CPUs don't have enough memory bandwidth left to run at full speed.

        • by ravyne ( 858869 )
          Unified memory isn't strictly bad, so long as there's sufficient bandwidth and sharing the resource isn't overly contentious. The PS4 has around 176GB/s bandwidth to its GDDR5. Xbox One has 68GB/s to its DDR3, but also has 32MB ESRAM that's effectively a software-controlled, level-4 cache shared between the CPU and GPU--it provides another 120GB/s bandwidth so Xbox One actually has a bit more bandwidth to go around than PS4 if the software is good about using the ESRAM. There's less apparent bandwidth to go
    • Maybe you didn't target your game properly.

      Many Nintendo games run in 1080p and 60 frames/second on the Wii U which is much less powerful...because Nintendo makes that their target when deciding how much AI and graphics detail to put on the screen at once.

      In other words, "cripple your game so it'll run on proprietary pieces of shit instead of letting it be as good as it possibly could be on a real computer."

      • by tepples ( 727027 )
        Unless "as good as it possibly could be" includes letting other players join in without each having to buy and bring another gaming PC.
    • "Many Nintendo games run in 1080p and 60 frames/second on the Wii U" What are these many you speak of? There's 1, Rayman, which BTW isn't even made by Nintendo, which is 1080p/60FPS on all the home consoles. Mario Kart 8 is 720p/59 output at 1080p. WW HD was Bayonetta 2 is 720p and SSB U isn't even out yet so who knows what it'll be. I played the Bayonetta 2 demo today and it looked freakin' awesome regardless.
  • by Sycraft-fu ( 314770 ) on Thursday October 09, 2014 @03:13PM (#48106063)

    Just saying if consoles aren't powerful enough to make you happy, well there are these new fangled PC things with a shit ton of CPU, RAM, and other goodies and gamers like me who spend way too much money on them to play games. Of course if you keep taking the attitude that we are all pirates, releasing shitty ports and so on don't be surprised if we aren't so interested in your products

    http://www.escapistmagazine.co... [escapistmagazine.com].

  • Do you mean to imply that chip manufacturers have started to lie on their spec sheets? Really? That's a new one; in the past they've been 100% accurate.

  • Cell (Score:4, Interesting)

    by SumDog ( 466607 ) on Thursday October 09, 2014 @03:14PM (#48106075) Homepage Journal

    I really hate that Sony dropped their cell processors going from the PS3 to the PS4 in favour of an x86 based system. We didn't see a lot of devices using cell and because of that, a lot of cell super-computer clusters were even made using actual PS3s. Even the prior MIPS processers of the earlier PlayStations are used in computer architecture texts books to this day (albeit overly-simplified versions of MIPS's pipling systems).

    I really want to see more architecture options, not less. Intel bought Alpha, killed it, screwed up with their own VLIW attempt with the Itaniums (which use EPIC) and I haven't heard anything about Transmeta in years. Today everything is ARM or x86_64 (with MIPS still seen in some embedded systems, mostly home routers). IBM still produces new POWER systems, but they're limited to a specific server niches.

    • Re:Cell (Score:5, Informative)

      by craighansen ( 744648 ) on Thursday October 09, 2014 @03:28PM (#48106253) Journal

      You haven't heard anything about Transmeta in years because they ceased operating in 2009. The patent portfolio went to Intellectual Ventures, LLC, and licensed in whole or part to Intel, Nvidia, Sony, Fujitsu, and NEC.

    • It was IBM's decision to end cell development due to it not having much of a market other than the PS3's and super-computers. Given the lackluster specs Sony went with for the PS4 it is unlikely that they could have convinced IBM to continue with cell's development even if they had tried.
    • Re:Cell (Score:5, Insightful)

      by ravyne ( 858869 ) on Thursday October 09, 2014 @03:40PM (#48106371)
      Cell -- at least the part of Cell that provided the computational grunt -- was not a CPU. Cell was a single-core, dual-threaded PowerPC core -- the exact same PowerPC core as the three in the Xbox360, save the extended vector instruction set and registers that Microsoft had added to their implementation. That core was basically the Intel Atom of PowerPC architectures. The better part of Cell was the DSP units, which you can basically think of as SIMD (Like SSE) units that are able to run autonomously from a small pool of local memory. The PowerPC's job is to load the Cell units with work (data and code) and coordinate the results -- they work for the CPU, and aren't capable of much on their own.

      In the PS4, you don't have cell, but instead you have 4 GPU cores dedicated for compute tasks (they don't have the back end necessary to perform rendering, although they can participate in compute tasks that aid the rendering pipeline.) Like Cell, these cores work for the CPU, have generally the same programming model (load them up with code and data and set them to running), and also have the exact same theoretical throughput as Cell had.

      Variety and competition are great, but Cell was nothing special in the sense that what was good and unique about it has been subsumed by GPU compute -- it was ahead of its time, but it hasn't aged nearly as well. Game consoles are a commodity business though, its hard to justify custom or obscure hardware unless its the only way to get the architecture you want, but then you have to teach everyone to use it effectively.
    • The SPARC processors are still doing some interesting things, and there's been some shakeups in the GPU architecture space.

    • Re:Cell (Score:5, Insightful)

      by AmiMoJo ( 196126 ) * on Thursday October 09, 2014 @04:01PM (#48106579) Homepage Journal

      The problem with the PS3's Cell and the PS2's odd set up before it is that they were both a bugger to get good performance from. It took developers years to get the best from them, and it cost a lot of money to do so. Compare that with systems like the Dreamcast and the two XBOX consoles that were relatively easy to get on with and which had excellent looking games from the start.

      With this generation both Sony and Microsoft realized that games are now so expensive to produce that anything which reduces that cost will be attractive. Most are cross platform too, so if you make your system hard to work with it's going to suffer from lame ports. The PS3 often had that problem, with its versions being inferior to the PC and 360 ones.

      Thus both new systems are basically PCs. Familiar hardware, familiar and fairly easy to work with CPUs, mature tools.

    • Re:Cell (Score:5, Interesting)

      by Dixie_Flatline ( 5077 ) <vincent@jan@goh.gmail@com> on Thursday October 09, 2014 @04:16PM (#48106705) Homepage

      Disclaimer: I work for Ubisoft. I did not work on the game in question and I won't comment on it.

      Now, the PS3. I have a friend that's made a very good living for the last few years doing nothing but PS3 optimisation. He'd go in 3 days a week and make more than I would in a year. The PS3 setup was fiendishly complicated and difficult to wring real performance out of. Even by the end of the cycle, I'd say there were only a few games that significantly made use of the potential power that was available in the PS3. On paper, it was impressive. In practice, it was a mild nightmare. You had completely different tools than when you were making a 360 game. The compiler was different. You had to be a lot more meticulous about where data was and how you were moving it around.

      I worked on the PS4 earlier this year, and it's dead easy to use. The tools integrate well into the environment, and you don't have nearly the same optimisation headaches that you did on the PS3. It's trivially faster than the XBone, and there's virtually no platform specific code (except for the obvious stuff, like connecting to the respective online services, etc.)

      From a developer perspective, the PS4 is a lot nicer than the PS3. That'll mean more simultaneous releases on the PS4 and XBone, and this time there's no delay before the PS4 is at or past parity with its competition (which is more important for Sony and Sony fans, really).

      That's just my opinion on the matter, but Sony really listened to the developer community when it came to tools and ease of use. It may be less interesting, but interesting generally means 'troublesome', not 'exciting' when you're writing software.

      • In that case, perhaps you know whether the APUs and the OS allow for general purpose computation? I can only speculate on whether the drivers allow for it, but if it's possible, chances are that substantial AI could be run on a few of the GPU cores. After all, modern AI is not just branchy Prolog engines.
      • by marsu_k ( 701360 )

        OK, so you work for Ubisoft, on to the topic at hand - Unity is not going to be on PS3 or Xbox 360. When we are talking about the current consoles (PS4 and Xbox One), they are basically the same - the difference being one has a slightly higher clocked CPU and a massively inferior GPU, along with much less memory bandwidth (yes, there's the "massive" ESRAM totaling a whopping 32MB), versus a slightly lower clocked CPU, much better GPU and a memory bandwidth in a totally different ballpark.

        Given these specs,

    • We didn't see a lot of devices using cell and because of that, a lot of cell super-computer clusters were even made using actual PS3s

      Sony sold and priced Cell based systems for commercial use.

      The PS3 could be purchased in wholesale lots --- taken out of consumer distribution channels where the true cost of the hardware would be recouped by future video game sales --- at a substantial net loss to Sony.

      Exit the "Other OS."

  • by Spy Handler ( 822350 ) on Thursday October 09, 2014 @03:16PM (#48106091) Homepage Journal

    quad core. Is it better than the PS/4 CPU?

    • by Fwipp ( 1473271 )

      Newegg has your processor at $240 - the entire PS4, including controller & game is only $400. I'd be curious to know what you shelled out for the full computer (please include input devices).

      You've always been able to buy more powerful machines than the consoles. Just not at console prices.

      • For the Xbox One...

        The most comparable processor is probably the AMD FX-8320, under-clocked. It'll cost you $150.
        The most comparable video card is probably the ADM 7770, also scaled back a touch, and it'll cost you $100.
        8GB of DDR3 will cost you $75
        The Blu-ray reader, another $40.
        500GB hard drive is another $40.

        That leaves you negative $5 for a motherboard, HTPC case, power supply and remote joypad.

        • It's rumored that PC's can do more than play games, though.

          What was the last time you posted on Slashdot through your Xbone?

          • My HTPC could post to Slashdot too, but it doesn't. It just records CATV and plays it (and downloaded videos) back for me with a high Wife-Acceptance-Factor interface.

            My current setup is an Xbox One with the "TV" input running to a Windows Media Center device with a Ceton cable cards tuner in it. I'm fairly angry that The One isn't a Media Center Extender, but I see Microsoft's vision of turning the One into that "one" device in the living room, and it means slowly, silently killing Media Center Edition.

            W

          • The Xbox 360 and One both have access to Internet Explorer, so that's not as far-fetched as you'd think.
          • by tepples ( 727027 )
            And when was the last time you plugged two to four USB controllers into your gaming PC or played a game in the "platform fighter" genre [tvtropes.org]?
        • Actually, the most comparable CPU would probably be a dual-socketed Athlon 5150 (or 5350) configuration (if there were MBs for that).
      • But you also need a computer, for school, work, facebook, whatever else, everybody has one nowadays. So the choice is between a $800 computer vs a $400 computer plus $400 console.

        • On the one hand, the $800 PC can run a far larger library of games. On the other, fewer of them tend to be in the sort of console-centric genres that cause people to buy one to three extra controllers for their consoles. Nor do JRPGs see wide release on anything but PlayStation platforms.
    • Yes.
  • That would explain frame rate, if they couldn't get their physics/AI routines to work fast enough (although they could use interpolation routines to alleviate that), but does not explain the rendering resolution. Do they expect us to believe they had the CPU working on shaders or rasterization? Even post-processing effects would be handled by the GPU.
  • boo hoo hoo (Score:5, Insightful)

    by cdrudge ( 68377 ) on Thursday October 09, 2014 @03:18PM (#48106121) Homepage

    Boo fricking hoo. Learn to develop a game with what you have and quit yer bitching.

    The best damn video game console ever was 8 bits, ran on a single core (usually), at 1.79 MHz IIRC. Gamers then logged just as many hours saving princesses, shooting aliens, and stacking blocks as what gamers do today. And guess what...they loved it. They fucking ate it up and went back for seconds and thirds. No, it wasn't photo-realistic 3D video with dolby-i-don't-give-a-crap sound and 87 button LCD-screen force feedback controllers. We didn't need it because we could have fun with what we had and didn't worry about frame rates or pixel resolutions or how many cores a OS management thread ran on vs graphics cores vs whatever.

    Game designers these days, spoiled rotten little twerps that whine about everything.

    • by Matheus ( 586080 )

      Get off my lawn!!!

    • Re:boo hoo hoo (Score:5, Informative)

      by neoritter ( 3021561 ) on Thursday October 09, 2014 @03:55PM (#48106519)

      And back in our days when we wanted to play with our friends we had to trudge 15 miles through 10 feet of snow going up hill both ways!

    • Re:boo hoo hoo (Score:5, Insightful)

      by pla ( 258480 ) on Thursday October 09, 2014 @04:28PM (#48106771) Journal
      Boo fricking hoo. Learn to develop a game with what you have and quit yer bitching.

      More to the point - When you have the luxury of coding for a very specific platform (ie, a gaming console with a known hardware configuration and known performance profile), you have no excuse for failing to adjust your resource demands accordingly. And if you just can't physically dial down the load enough to run well on platform X - You don't release the goddamned game for platform X.

      Re-read that last point, because it nicely translates Pontbriand's whining into plain English: "We promise not to turn down any chance to grab your cash, no matter how shitty the experience for our loyal customers".
      • I have a radical idea. If we accept that game developers are going to have the moral equivalency of a used car salesman with a crack addict to support, how about we ( the customers ) stop giving said addicts OUR GOD DAMNED MONEY FOR SHIT.

        Anyone who gives these greedy SOBs a single cent are part of the problem.

    • I should know; I've programmed for that platform [pineight.com]. But there are a few kinds of game that can't easily be adapted to it [nesdev.com]. One is games where you need to see farther away than 4 to 8 player heights. Those typically need a 3D GPU to draw a behind-the-player perspective.
  • by craighansen ( 744648 ) on Thursday October 09, 2014 @03:19PM (#48106131) Journal

    TFA just mentions the interview without a clear reference to it. Looking for it I found two other articles that suggest that the 900p resolution and 30fps targets came from other factors. http://www.gamespot.com/articl... [gamespot.com] says that 30fps is "more cinematic" and 60fps "looked really wierd." http://www.gamespot.com/articl... [gamespot.com] suggests that some non-graphic computation is going on the GPU, but also has a quote that mentions "technically CPU bound."

    What we don't know from these articles is why some or more of the AI computation can't be done in the GPU.

    • What we don't know from these articles is why some or more of the AI computation can't be done in the GPU.

      Because modern GPUs doesn't have a 1:1 mapping between their "cores" and general purpose logic units. Once you use a branch you effectively collapse your "cores" down to the general purpose logic associated with those "cores." (I think my gtx 660ti has like ~1400 cores and ... 8 general purpose logic units?) For graphics you essentially are just doing vector/matrix math calculations with no branching so you can use all of the "cores" in parallel. This is not to mention that there are costs in streaming d

      • What makes you certain that the AI code can't be vectorized across the GPU cores, for example, one core per NPC? Yes, we understand that CUDA/GPU's don't run vanilla C code.

      • GCN cores were specifically designed to expand the range of code suitable for running on the GPU. Among other things, they should be able to access native data structures used by typical CPU code with little or no overhead, requiring no data copies. That's the whole point of the HSA approach. And quite a lot of AI stuff could probably be run on them. Now the question is how friendly the PS4's specific firmware/OS is to this style of programming.
        • by Fwipp ( 1473271 )

          Yep, this is totally right. The main thing keeping an algorithm from running well on GCN cores is being branch-heavy. While I haven't kept up-to-date on the terminology, only one instruction at a time can be executing across a set of inputs.

          The following code is pretty understandable and quick on a CPU. But on a GPU, performance suffers.
          if (unitOnFire){
          flailAround();
          } else {
          doFightRoutine();

  • I'd go for reducing overall visual quality rather than the AI quality. Having good AI, and supposedly better gameplay as a result, should be at the top of the priority list. I'd go so far as to reduce the visual quality so much that the framerate is high and the game responsiveness is super smooth.

  • F Ubisoft (Score:2, Insightful)

    by Anonymous Coward

    Why should be believe anything he has to say? He is a known employee of Ubisoft, after all.

  • by medv4380 ( 1604309 ) on Thursday October 09, 2014 @03:21PM (#48106161)
    I suspect BS, but I'll hear out the argument if there is actual evidence. Sliding down the resolution to 900p from 1080p would mostly save you on GPU, and Graphics memory usages. In a more detailed artilce it was stated that they picked 900p because they didn't want to fight with the differences between the XBone and PS4. The main reason the PS4 keeps getting 1080p and the XBone does not is that the memory for the graphics is so much faster. So if someone actually has the PC version and can so that "AI" is burning every processor at 100% then I'll buy their argument as plausable. Until then the story has changed from day to day, and isn't believable. If the AI has something to do with it then the AI coder has probably deadlocked the system when they use multiple cores, and has been cheating by using only one core.
    • The PS4 has GDDR5 RAM while the Xbox-1 uses GDD3 RAM with 32MB of high-speed ESRAM integrated directly into to processor die. This means that Xbox-1 code needs design considerations and optimized code to take advantage of the small but super-high speed cache RAM. Most cross-platform developers simply will not spend resources on that.
    • by EvilSS ( 557649 )

      I suspect BS, but I'll hear out the argument if there is actual evidence. Sliding down the resolution to 900p from 1080p would mostly save you on GPU, and Graphics memory usages. In a more detailed artilce it was stated that they picked 900p because they didn't want to fight with the differences between the XBone and PS4. The main reason the PS4 keeps getting 1080p and the XBone does not is that the memory for the graphics is so much faster. So if someone actually has the PC version and can so that "AI" is burning every processor at 100% then I'll buy their argument as plausable. Until then the story has changed from day to day, and isn't believable. If the AI has something to do with it then the AI coder has probably deadlocked the system when they use multiple cores, and has been cheating by using only one core.

      Well if Watch_Dogs (another UbiSoft title) is any indication then I would believe it. I bought the PC version (yea yea, I know, Watch_dogs!? Fool me once...) and ran it on a 2nd gen i5. Don't recall the model exactly but it was clocked north of 3Ghz. Not a beast by today's standards but not horrible either. With that and a 780ti, the game was horribly CPU bound. Constantly at 100% CPU. Upgrading to a i7 4790K and it still runs at around 60% CPU. So yes, I can completely believe that their games are CP

  • How do we know that the real issue isn't scripting silliness and otherwise inefficient code design? Not saying this is the case, just pointing out that the possibility is glossed over.

    • That's much more likely to be the case than whatever nonsense some PR person is feeding these journalists. Ubisoft isn't known for their optimization prowess.

  • As an embedded architect/programmer, I deal with this all of the time. You have to design for the platform, not what you fantasize about having. Doing the dev work on a dual-4-core-hyperthreaded box with 48 GBytes of memory, then whining that the CPU/GPU/memory space/... isn't enough is just exposing your stupidity. Onbce had a developer build and test a database on his quad-core Apple-thingy, then whine to management that the hardware engineers and materials people had just not done their job, when, eve

  • by Anonymous Coward on Thursday October 09, 2014 @03:41PM (#48106387)

    ... i thought they were going to blame it on piracy?

  • are you saying this gen isn't much better then last gen omg the realty.

Avoid strange women and temporary variables.

Working...