Sticky

Nvidia Volta Megathread

Seems Volta is not really discussed much here, besides being mentioned as a Vega competitor. Time to get some info on the board?

https://www.extremetech.com/extreme/249106-nvidia-goes-ai-new-volta-architecture

This article gets me thing that Volta is not just GPU but there is an AI and general computing side as well. May sound far fetched but at may have repercussions for us. If AI is now becoming a mainstream thing in business and special applications, we might see consumer applications some five years down the road. Especially considering that players like NVidia are willing to open source their code. I mean some of you guys might end up using your Volta GPUs for AI on your rigs, eg. things like truly learning AI in games. But I think AI will have a broader application than just gaming - even on the local PC. Not to mention there are already some applications that use the GPU computing power for special applications that are not AI.

Long-term, it seems there is a lot of hype around AI currently and certainly it looks like self-driving vehicles are just around the corner. It reminds me of the 1960s when human level AI was supposed to be just around the corner. But it's quite possible that in the first half of 21st century we'll see a number of mainstream AI applications and it's going to be one of the next big technological waves. (We're still waiting for a real biotech wave too.)
Reply to varis
130 answers Last reply
More about true volta thread
  1. nvidia definitely going all out with AI. and they will not going to give up the market completely to ASIC hence why they make their own Tensor Core inside their GPU. many people often only look nvidia as graphic company only. hence many expect that once nvidia secure their lead in GPU performance they will start to have it easy and start charging expensive price for very minimal performance improvement like intel. but nvidia really have no time to rest easy. because they also competing with other company like google in the field of AI for example. they were able to make their architecture more power efficient with kepler but that is not enough because in mobile (where they compete with tegra) there are more power efficient architecture exist. while their venture into smartphone and tablet did not go as plan they still benefit a lot from their experience dealing with mobile. that's why even if nvidia already ahead of AMD in GPU they still have to work harder.
    Reply to renz496
  2. AMD and NVidia have announced their press conferences at Computex 2017: NVidia is tomorrow tuesday and AMD coming wednesday.

    https://techdrake.com/2017/05/nvidia-and-amd-announced-computex-2017-press-conferences/

    Overview of what's coming in the conf:

    http://www.trustedreviews.com/news/computex-2017-news-launches-nvidia-intel-amd-asus-microsoft-announcements

    Some reporters seem to even think NVidia would launch Volta at Computex. I'd hope for GTX20 at best...

    Liveblog:

    http://www.anandtech.com/show/11457/asus-computex-2017-press-conference-live-blog
    Reply to varis
  3. honestly i don't think we will going to see any volta related today. meanwhile board partner are gearing up to showcase some of their extreme 1080ti for computex. that effort will be useless if nvidia suddenly release 1080ti successor today

    https://videocardz.com/69791/msi-to-unveil-geforce-gtx-1080-ti-lightning-z-at-computex
    https://videocardz.com/newz/evga-teases-upcoming-products
    https://videocardz.com/newz/colorful-igame-gtx-1080-ti-neptune-pictured
    Reply to renz496
  4. Sane thinking. So far Volta was discussed but it was all about AI and tensor cores.

    Lighter gaming laptops with geforce are coming though :)
    Reply to varis
  5. the main star on the gaming side of the presentation is no doubt was the new thin gaming laptop. JHH even show actual product on stage. that's is like to show off that you can get such performance on laptop right now where as there isn't something similar from "competitor".
    Reply to renz496
  6. I think there are 2 main takeaways for the "Volta generation" from nVidia and they won't be pure "GPU" enhancements. There will be that as well, but my take is nVLink for consumer and hardware schedulers for nVidia. Those two are really interesting to me at least, even if "brute performance" doesn't go up two-fold.

    Cheers!
    Reply to Yuka
  7. the 1080ti just released, we wont see its successor for quite some time, probably a year and a half. a side grade 1170/1180 may show up, but big volta wont be around for a long time.
    Reply to nikoli707
  8. So what exactly does AI incorporation into Volta architecture mean for gaming? I mean it's badass but I'm having issues trying to imagine the impact it will make in the pc gaming scene.
    Reply to techbard
  9. techbard said:
    So what exactly does AI incorporation into Volta architecture mean for gaming? I mean it's badass but I'm having issues trying to imagine the impact it will make in the pc gaming scene.


    probably none just like how it was with pascal. though realistically FP16 can still somehow benefit gaming (this is what AMD try to do with Vega). but those tensor cores? honest i don't know. that thing is pretty much specific function designed for AI related task. but then again GV100 is designed in a way to be a compute monster. simultaneous FP32 and INT32 operation. massive FP64 (1:2 FP32 ratio), massive FP16 and the new tensor core that are more optimized for AI specific task. at this point we just don't know if those tensor core is "mandatory" to all volta design or specific to GV100 only.
    Reply to renz496
  10. so nvidia indeed researching their own infinity fabric for their future GPU?:

    http://techreport.com/news/32189/nvidia-explores-ways-of-cramming-many-gpus-onto-one-package
    Reply to renz496
  11. In terms of market, Nvidia has established a generational lead over AMD (the latest AMD gaming cards are only as fast as Nvidia's second best, two year old cards). With the release of Vega, I'm looking for Nvidia to make some Volta news towards the holiday buying season. Not that they will have a new product to sell, but that they can insert some doubt into anyone looking to purchase the current generations. I think they will make maintaining that lead an important strategy.
    Reply to 17seconds
  12. Well, they already said: "hey, look, we'll park Volta release for a few extra months, due to... reasons; but hey! there's plenty of Pascal stuff still!".

    Cheers!
    Reply to Yuka
  13. well if we look at nvidia past release it is more realistic to expect Volta in Q2 2018. earliest in Q1 next year. for sure no gaming volta this year since nvidia already said so during their latest earning.
    Reply to renz496
  14. renz496 said:
    well if we look at nvidia past release it is more realistic to expect Volta in Q2 2018. earliest in Q1 next year. for sure no gaming volta this year since nvidia already said so during their latest earning.


    Oh, yes. I meant that in the gaming-plebeian context. I don't believe they'll have delays for the "pro" stuff..

    Cheers!
    Reply to Yuka
  15. Yuka said:
    renz496 said:
    well if we look at nvidia past release it is more realistic to expect Volta in Q2 2018. earliest in Q1 next year. for sure no gaming volta this year since nvidia already said so during their latest earning.


    Oh, yes. I meant that in the gaming-plebeian context. I don't believe they'll have delays for the "pro" stuff..

    Cheers!


    actually i did not expect even the "pro" volta to be ready in late 2018 or Q1 2018. but the competition in AI space is heating up. nvidia need to make sure GPU not to be out done by chip that is build specifically for AI application only like google TPU. those that pre-ordered DGX-1 before nvdia officially announced tesla V100 being upgraded to tesla V100 without additional charge. it seems nvidia want to replace tesla P100 completely. that is very fast even by nvidia standard. just look how long nvidia keep their GK110 as their compute flagship before they replace them with P100.
    Reply to renz496
  16. I knew you guys would help me resurrect this thread.
    Reply to 17seconds
  17. nothing much to discuss on volta right now. especially on the gaming side of things. but on compute side of things there probably some new enhancement that will make volta a monster in mining......
    Reply to renz496
  18. Uhm... I'll go on a tangent here, because I don't really know the inner workings on AI development nor there is abundant information on implementation and design for it.

    From the holistic design perspective, Artificial Intelligence has 2 main aspects to it: Process and Analysis.

    For Process, I'm putting in the same bag the "simulation" or adaptation of thought process into a machine lang (Lisp or any lang that is actually used) where you can indeed tailor ASIC stuff without much complexity in the process-to-metal translation.

    For Analysis, I'm putting in the same bag all the "computation" behind each singular point of decision and intake of information given (massively parallel is an understatement here). This is a huge bag, even bigger than process, but straight forward: you can process as much as you can calculate (input vs output to an effect), so the effective decision nodes will be a reflect of that amount of data you process in time (kind of obvious, right?). Simple in concept, a nightmare in actual implementation (those trees!) since coordinating the decisions into a single point of action is quite the challenge, depending on implementation.

    So, what I'm trying to get at here is the point of nVidia being competitive in this AI landscape.

    Much like you can use multipurpose CPUs to do whatever you want, you can use GPU for "general purpose" training and designing of AIs with some degree of "openness" in approaches. Specific design should come after you have chosen a certain way of designing your AI process, I would imagine and the rest is just growing on the capabilities of the analysis you want? I would imagine since the computational needs, at the end of the day, will be tailored for mass production (specific products with "AI" fed operations), you would choose that method for mass, but interim would always be "general"; or I would imagine it's the feasible way. All of that to say that I do think nVidia has a chance to grab a lot of market as long as they can improve on the analysis side as much as they can. Hence nVLink and all I/O improvements they can make to "feed the beast" will become a deciding factor to their success, in my opinion, obviously. Some of those can trickle down to consumers, I would imagine, but I don't really see how the general purpose calculation route will help "push more triangles" for games. At the end, it seems nVidia is better of making specific AI-accelerators like Google and if Volta is that first step, I won't expect huge leaps in the gaming arena TBH.

    Cheers!
    Reply to Yuka
  19. it all depends on what kind of improvement nvidia will come up with gaming volta. right now i'm suspecting nvidia will increase Volta gaming performance using the similar method they able to gain more performance with maxwell even when maxwell was build on the same 28nm process as kepler; architecture re-work. while volta will be build on 12nm from what i heard it was just some fancy name for more optimized 16nm process. GV100 is a lot bigger and faster than GP100 but power consumption wise both are about the same.
    Reply to renz496
  20. Even if the full version of GV102 comes in with 4096 cores, I'm more interested in at what clock and power numbers. Right now, I'm thinking the GV102 needs to hit 2.5 GHz out of the box and reach 3 GHz when overclocked. It needs to do both things on air with a similar power budget to GP102. A large shortfall on any one of these targets would be a disappointment to me. They also need GDDR6 to deliver it's target performance on day one. Greedy of me, I know. Maybe unrealistic. I hope not.
    Reply to manleysteele
  21. manleysteele said:
    Even if the full version of GV102 comes in with 4096 cores, I'm more interested in at what clock and power numbers. Right now, I'm thinking the GV102 needs to hit 2.5 GHz out of the box and reach 3 GHz when overclocked. It needs to do both things on air with a similar power budget to GP102. A large shortfall on any one of these targets would be a disappointment to me. They also need GDDR6 to deliver it's target performance on day one. Greedy of me, I know. Maybe unrealistic. I hope not.


    rather than increasing the clock i think it is better for nvidia to improve the IPC of their architecture instead. increasing the clock further might not going to increase the performance that much. we already seeing the diminishing return with pascal. as for what memory they used i don't care much about it as long as they can get the performance target.
    Reply to renz496
  22. renz496 said:
    manleysteele said:
    Even if the full version of GV102 comes in with 4096 cores, I'm more interested in at what clock and power numbers. Right now, I'm thinking the GV102 needs to hit 2.5 GHz out of the box and reach 3 GHz when overclocked. It needs to do both things on air with a similar power budget to GP102. A large shortfall on any one of these targets would be a disappointment to me. They also need GDDR6 to deliver it's target performance on day one. Greedy of me, I know. Maybe unrealistic. I hope not.


    rather than increasing the clock i think it is better for nvidia to improve the IPC of their architecture instead. increasing the clock further might not going to increase the performance that much. we already seeing the diminishing return with pascal. as for what memory they used i don't care much about it as long as they can get the performance target.


    Diminishing returns from Pascal? You're kidding me, right? Pascal buries Maxwell at the same core count and that burying is almost all due to clock. Plus it reaches those higher clocks using less power.
    Reply to manleysteele
  23. manleysteele said:
    renz496 said:
    manleysteele said:
    Even if the full version of GV102 comes in with 4096 cores, I'm more interested in at what clock and power numbers. Right now, I'm thinking the GV102 needs to hit 2.5 GHz out of the box and reach 3 GHz when overclocked. It needs to do both things on air with a similar power budget to GP102. A large shortfall on any one of these targets would be a disappointment to me. They also need GDDR6 to deliver it's target performance on day one. Greedy of me, I know. Maybe unrealistic. I hope not.


    rather than increasing the clock i think it is better for nvidia to improve the IPC of their architecture instead. increasing the clock further might not going to increase the performance that much. we already seeing the diminishing return with pascal. as for what memory they used i don't care much about it as long as they can get the performance target.


    Diminishing returns from Pascal? You're kidding me, right? Pascal buries Maxwell at the same core count and that burying is almost all due to clock. Plus it reaches those higher clocks using less power.


    IPC wise both maxwell and pascal are identical. true pascal can be clocked higher but we have to remember architecture wise both masxwell and gaming pascal is almost outright identical. when we increase the core clocked performance increase will not going to be linear. there will be point where the increase of clock will net much smaller performance increase. also heat can be a problem at high frequency. but if you think pascal is amazing then volta probably will be more frightening lol.
    Reply to renz496
  24. renz496 said:
    manleysteele said:
    renz496 said:
    manleysteele said:
    Even if the full version of GV102 comes in with 4096 cores, I'm more interested in at what clock and power numbers. Right now, I'm thinking the GV102 needs to hit 2.5 GHz out of the box and reach 3 GHz when overclocked. It needs to do both things on air with a similar power budget to GP102. A large shortfall on any one of these targets would be a disappointment to me. They also need GDDR6 to deliver it's target performance on day one. Greedy of me, I know. Maybe unrealistic. I hope not.


    rather than increasing the clock i think it is better for nvidia to improve the IPC of their architecture instead. increasing the clock further might not going to increase the performance that much. we already seeing the diminishing return with pascal. as for what memory they used i don't care much about it as long as they can get the performance target.


    Diminishing returns from Pascal? You're kidding me, right? Pascal buries Maxwell at the same core count and that burying is almost all due to clock. Plus it reaches those higher clocks using less power.


    IPC wise both maxwell and pascal are identical. true pascal can be clocked higher but we have to remember architecture wise both masxwell and gaming pascal is almost outright identical. when we increase the core clocked performance increase will not going to be linear. there will be point where the increase of clock will net much smaller performance increase. also heat can be a problem at high frequency. but if you think pascal is amazing then volta probably will be more frightening lol.


    Your first sentence confirmed my point, nicely. On your second assertion, perhaps, but at what level of clock does this occur? We're not there yet. Heat is an ongoing problem, even at present frequencies and densities. I do expect gaming Volta make some tradeoffs between power, clock, and core count. Still, I also expect a good bump on the full performance metric, both in the middle and on the high end.
    Reply to manleysteele
  25. Quote:
    but at what level of clock does this occur?


    maybe somewhere above 2.1ghz mark? probably why nvidia decided to "lock" pascal at pretty much 2.0-2.1Ghz limit right now. they can probably make work around to make it clock higher but will the effort going to worth it? just look at AMD Vega. AMD spend 3.9 billion of transistor just so GCN can be clocked much higher. they finally able to do it but their power consumption also going crazy.
    Reply to renz496
  26. renz496 said:
    manleysteele said:
    renz496 said:
    manleysteele said:
    Even if the full version of GV102 comes in with 4096 cores, I'm more interested in at what clock and power numbers. Right now, I'm thinking the GV102 needs to hit 2.5 GHz out of the box and reach 3 GHz when overclocked. It needs to do both things on air with a similar power budget to GP102. A large shortfall on any one of these targets would be a disappointment to me. They also need GDDR6 to deliver it's target performance on day one. Greedy of me, I know. Maybe unrealistic. I hope not.


    rather than increasing the clock i think it is better for nvidia to improve the IPC of their architecture instead. increasing the clock further might not going to increase the performance that much. we already seeing the diminishing return with pascal. as for what memory they used i don't care much about it as long as they can get the performance target.


    Diminishing returns from Pascal? You're kidding me, right? Pascal buries Maxwell at the same core count and that burying is almost all due to clock. Plus it reaches those higher clocks using less power.


    IPC wise both maxwell and pascal are identical. true pascal can be clocked higher but we have to remember architecture wise both masxwell and gaming pascal is almost outright identical. when we increase the core clocked performance increase will not going to be linear. there will be point where the increase of clock will net much smaller performance increase. also heat can be a problem at high frequency. but if you think pascal is amazing then volta probably will be more frightening lol.


    ipc maxwell considerably faster than pascal. a 980ti@1600mhz matches a 1070@2200mhz. and kingpins 2000mhz+ 980ti world record at the time was something like 27000ish in firestrike gpu score which is close to a 1080ti in stock form. all that really does mean much, we just want 20-40% real world gaming performance increases at the same power draw. i dont really care what the clock rate or memory config is, just give me performance.
    Reply to nikoli707
  27. nikoli707 said:
    renz496 said:
    manleysteele said:
    renz496 said:
    manleysteele said:
    Even if the full version of GV102 comes in with 4096 cores, I'm more interested in at what clock and power numbers. Right now, I'm thinking the GV102 needs to hit 2.5 GHz out of the box and reach 3 GHz when overclocked. It needs to do both things on air with a similar power budget to GP102. A large shortfall on any one of these targets would be a disappointment to me. They also need GDDR6 to deliver it's target performance on day one. Greedy of me, I know. Maybe unrealistic. I hope not.


    rather than increasing the clock i think it is better for nvidia to improve the IPC of their architecture instead. increasing the clock further might not going to increase the performance that much. we already seeing the diminishing return with pascal. as for what memory they used i don't care much about it as long as they can get the performance target.


    Diminishing returns from Pascal? You're kidding me, right? Pascal buries Maxwell at the same core count and that burying is almost all due to clock. Plus it reaches those higher clocks using less power.


    IPC wise both maxwell and pascal are identical. true pascal can be clocked higher but we have to remember architecture wise both masxwell and gaming pascal is almost outright identical. when we increase the core clocked performance increase will not going to be linear. there will be point where the increase of clock will net much smaller performance increase. also heat can be a problem at high frequency. but if you think pascal is amazing then volta probably will be more frightening lol.


    ipc maxwell considerably faster than pascal. a 980ti@1600mhz matches a 1070@2200mhz. and kingpins 2000mhz+ 980ti world record at the time was something like 27000ish in firestrike gpu score which is close to a 1080ti in stock form. all that really does mean much, we just want 20-40% real world gaming performance increases at the same power draw. i dont really care what the clock rate or memory config is, just give me performance.


    A 980Ti has 2816 cores. A 1070 has 1920 cores. I agree that performance/watt is what we are looking for. The comparison you are looking for is the 2048 core GTX 980 vs. the 1920 core GTX1070. Still not exactly equivalent, but life is unfair.
    Reply to manleysteele
  28. Most often people only look at core clock but overlook the core count. On architecture level maxwell and pascal structure is almost identically the same. So IPC wise it ahould be identical (per anandtech assumption). Just look at GTX980 vs GTX1060. GTX1060 need much higher clock to reach 980 performance but 1060 also have much less CUDA core than 980 (1280 vs 2048).
    Reply to renz496
  29. I have one problem with looking at individual "execution units" (or shader arrangements) and taking them as part of the "efficiency" equation.

    That is similar to taking ALUs, AGUs, FPUs and other internal execution units from a CPU and using them as an effective measure of "efficiency" in a CPU design. I know you *can* do it, but it's hard to justify it in context. When you count them, you're leaving other effective parts of the GPU that need to be included, such as memory controllers and... I can't think of another, haha.

    It's simpler to just analyze GPUs as a monolithic entity, much like we see a CPU core. GPUs don't have a divide (yet, they might actually go "MCM" shortly).

    Cheers!
    Reply to Yuka
  30. manleysteele said:
    nikoli707 said:
    renz496 said:
    manleysteele said:
    renz496 said:
    manleysteele said:
    Even if the full version of GV102 comes in with 4096 cores, I'm more interested in at what clock and power numbers. Right now, I'm thinking the GV102 needs to hit 2.5 GHz out of the box and reach 3 GHz when overclocked. It needs to do both things on air with a similar power budget to GP102. A large shortfall on any one of these targets would be a disappointment to me. They also need GDDR6 to deliver it's target performance on day one. Greedy of me, I know. Maybe unrealistic. I hope not.


    rather than increasing the clock i think it is better for nvidia to improve the IPC of their architecture instead. increasing the clock further might not going to increase the performance that much. we already seeing the diminishing return with pascal. as for what memory they used i don't care much about it as long as they can get the performance target.


    Diminishing returns from Pascal? You're kidding me, right? Pascal buries Maxwell at the same core count and that burying is almost all due to clock. Plus it reaches those higher clocks using less power.


    IPC wise both maxwell and pascal are identical. true pascal can be clocked higher but we have to remember architecture wise both masxwell and gaming pascal is almost outright identical. when we increase the core clocked performance increase will not going to be linear. there will be point where the increase of clock will net much smaller performance increase. also heat can be a problem at high frequency. but if you think pascal is amazing then volta probably will be more frightening lol.


    ipc maxwell considerably faster than pascal. a 980ti@1600mhz matches a 1070@2200mhz. and kingpins 2000mhz+ 980ti world record at the time was something like 27000ish in firestrike gpu score which is close to a 1080ti in stock form. all that really does mean much, we just want 20-40% real world gaming performance increases at the same power draw. i dont really care what the clock rate or memory config is, just give me performance.


    A 980Ti has 2816 cores. A 1070 has 1920 cores. I agree that performance/watt is what we are looking for. The comparison you are looking for is the 2048 core GTX 980 vs. the 1920 core GTX1070. Still not exactly equivalent, but life is unfair.


    Another way to look at it is that the 980 ti has 8 billion transistors, while the 1070 has 10% less, 7.2 billion.
    Reply to TMTOWTSAC
  31. World’s Largest Server Companies Announce NVIDIA Volta Systems Supercharged for AI

    https://www.hpcwire.com/off-the-wire/worlds-largest-server-companies-announce-nvidia-volta-systems-supercharged-ai/
    Reply to Hellfire13
  32. Nvidia is very agressive to maintain the lead they have on AI. They rely 4 years on GK110 for their top compute card before. Now GP100 has only been a year on the market and they already replace them with GV100 ( though nvidia still selling even GK110 based accelerator to the market).
    Reply to renz496
  33. There's one caveat though. From what I'm gathering, nVidia is pushing to lead the conversations for AI work in terms of technology. All the IP they are showing and discussing is proprietary stuff with zero hints of FOSS. This smells like a rehash of the CUDA strategy. I'm not saying it's bad or good, but that is what it smells like to my nose, heh.

    Cheers!
    Reply to Yuka
  34. Because for over a decade it has been working very well for them. Open source might be the best option in ideal world but the thing with Open source is everyone have a say in shaping the API/software. That's good so the software will work on all hardware but sometimes they can hold back things from going forward. I think that's why AMD very reluctant to give IHV direct access to their Mantle API before. They will give the base spec to Khronos group but they did not want nvidia or intel to have any hands in developing mantle itself.
    Reply to renz496
  35. Fair point.

    That is the good side of having a proprietary API in use: tailored performance. Open Source APIs have the "fit all" and the "backwards compatibility" problems. Those two are not easy nor simple to approach.

    Hence why I don't consider it a bad thing inherently. It's just... I get itchy when I have to deal with proprietary stuff, haha.

    Cheers!
    Reply to Yuka
  36. I think nvidia quite openly said they did not like having to do all the job while others just get a free rides. That's why they go with more proprietary route and if possible charge money for them. Consumer might not like this but this is probably why nvidia are in much better position than amd financial wise.

    Personally i would like nvidia to adopt more open standard (like adaptive sync) instead of seeing them as competitor to their solution. Though if nvidia indeed suporting adaptive sync it might end up hurting AMD more. To me despite the complain many people had for nvidia not supporting adaptive sync it give AMD some breathing room. Rigt now supporting adaptive sync becoming one of AMD advantage. Just look the recent pricing with Vega. It probably much less of a mess for AMD if nvidia decided to capitalize their lead by keeping 1080 original pricing and charge more for 1080ti.
    Reply to renz496
  37. It's just an approach they have that I (and a lot more people) don't really share nor like.

    FOSS, or free to use software, does not mean people is stealing things from other people when they contribute. The whole point is that "everybody wins". The take nVidia uses is to justify their own angle, but at least they're honest about it: they don't like sharing because they feel like they lose instead. That is fine; they're not hiding it either. They're a company after all. This model brings them the most profit, otherwise they'd use another business model.

    And I can give you a lot of examples where the FOSS is just as good as the private/closed version. And in terms of vulnerabilities, FOSS is better hands down.

    BUT! This is a very dense and diverse topic that is too tangencial to this thread, haha.

    In regards to nVidia having success, that is thanks to their engineers first and marketing team second, haha. AMD derping is also helping them a lot.

    Cheers!
    Reply to Yuka
  38. When do you think will I be able to get a notebook with a volta's replacement for 1070/1080 mobile? Want to know if it's worth the wait really,considering I don't really need a fast conputer anytime soon and I'm already getting a halfdecent 1080p machine with a rx560 (decent, lowpower uograde from gtx760) and a r5 1600 (from a8 6600k, which is slower than g4560). I have loads of older games ro play still and at medium settings I could play most newer games for the time being, especially since I'm finally uograding to freesync. However, I can't always use the desktop, hence why I'm not going all-out, and want to get a clevo coffee lake notebook with a 1080 or whatever that will last me as long as possible for gaming whereever in the house, so I don't have to wait for me sister to finish her homework.
    Reply to thegamingsteamtrain
  39. waiting can be worth it....but it can also be other wise. like those that waiting for Vega and hoping it will compete head to head with nvidia 1080ti. though notebook with nvidia next generation GPU probably won't be here anytime soon. desktop card will be here first. for the mobile version it might not coming out until a few months later.
    Reply to renz496
  40. renz496 said:


    I read that, too. My first thought was that is right on schedule for holding up a piece of silicon, saying we are introducing Ampere, giving a list of features, and then, saying it's the next gen after Volta. Available at some time in 2019. He's been doing this at GTC for years.
    Reply to manleysteele
  41. manleysteele said:
    renz496 said:


    I read that, too. My first thought was that is right on schedule for holding up a piece of silicon, saying we are introducing Ampere, giving a list of features, and then, saying it's the next gen after Volta. Available at some time in 2019. He's been doing this at GTC for years.


    reading the article it is more like Ampere will directly replace "gaming" pascal and will launch in Q2 2018. probably nvidia decided to use separate architecture name for their gaming and compute GPU. with pascal we have gaming and compute pascal. despite sharing the same architecture name compute pascal was build differently. with kepler even when GK110 were more compute focus than it's smaller sibling they still share similar design. for example all SMX in kepler have 192 cuda cores. with pascal "gaming" pascal have 128 cuda cores per SM. with GP100 each SM only have 64 Cuda cores each. anyway if this is real we should hear more early next year.
    Reply to renz496
  42. Interesting. Now i'm curious as to what this "Ampere" will provide.
    Reply to TechyInAZ
  43. Nvidia Is Building Its Own AI Supercomputer
    by Paul Alcorn November 13, 2017 at 3:00 PM

    Quote:

    AI has begun to take over the data center and HPC applications, as a quick trip through the recent HotChips and ongoing Supercomputing 2017 trade shows quickly affirms.
    Nvidia has been one of the companies at the forefront of AI development, and its heavy investments in hardware, and more importantly the decade-long development of CUDA, have paid off tremendously. In fact, the company announced that 34 new Nvidia GPU-accelerated supercomputers were added to the TOP500 list of supercomputers, bringing its total to 87. The company also powers 14 of the top 20 energy-efficiency supercomputers on the Green500 list.
    Jensen Huang, Nvidia CEO, took to the stage at Supercomputing 2017 to announce that the company is also developing its own AI supercomputer for its autonomous driving program. Nvidia projects its new supercomputer will land in the top 10 list of worldwide AI supercomputers, which we'll cover shortly.

    New threats loom in the form of Intel's recent announcement that it's forming a new Core and Visual Computing business group, headed by Raja Koduri of AMD fame, to develop not only a discrete GPU but also a differentiated graphics stack that scales from the edge to the data center. AMD is also enjoying significant uptake of its EPYC platform paired with its competitive Radeon Instinct GPUs, a big theme here at the show, while a host of FPGA and ASIC vendors are also vying for a slice of the exploding AI segment.
    The show floor is packed with Volta V100 demonstrations from multiple vendors, as we'll get to in the coming days, and Nvidia also outlined its progress on multiple AI fronts.
    Nvidia's SaturnV AI Supercomputer




    Quote:

    Nvidia also announced that it's upgrading its own supercomputer, the SaturnV, with 660 Nvidia DGX-1 nodes (more detail on the nodes here), spread over five rows. Each node houses eight Tesla V100s for a total of 5,280 GPUs. That powers up to 660 PetaFLOPS of FP16 performance (nearly an ExaFLOPS) and a peak of 40 PetaFLOPS of FP64. The company plans to use SaturnV for its own autonomous vehicle development programs.
    Amazingly, Nvidia claimed that the SaturnV will easily land in the top 10 AI supercomputers worldwide, and might even enter the top five once completed.
    That's impressive considering that state-run institutions drive many of the world’s leading supercomputers. Nvidia and IBM are also developing the Summit supercomputer at the Oakridge National Labs. Summit should unseat the current top supercomputer, which resides in China, with up to three ExaFLOPS of AI performance.
    Expanding The Ecosystem




    Quote:

    Nvidia announced that its massive V100 GPU, which boasts an 815mm2 die packed with 21 billion transistors and paired with 16GB of HBM2, has been adopted by every major system and cloud vendor. The blue-chip systems roster includes Dell EMC, HPE, Huawei, IBM, Lenovo, and a host of whitebox server vendors. AI development is expanding rapidly, and a number of startups, businesses, and academic institutions are rapidly developing new products and capabilities, but purchasing and managing the requisite infrastructure can hinder adoption.
    Every major cloud now offers Volta-powered cloud instances, including AWS, Azure, Google Cloud, Alibaba, Tencent, and others. Microsoft recently announced its new cloud-based Volta services, which will be available in November. The company also now supports up to 500 applications with the CUDA framework.










    Quote:

    Nvidia is also offering new software and tools on the Nvidia GPU Cloud (NGC), such as new HPC containers that simplify deployment. The containers work with Docker, with support for other container applications coming in the future. The applications run on any Pascal or newer GPU, along with HPC supercomputing clusters and Nvidia's DGX systems. These containers support all the main frameworks, such as Caffee, Torch, Pytorch, and many others. Nvidia has focused on simplifying to process down to a few minutes.
    Japan's AIST also incorporated 4,352 Tesla Volta V100 GPUs into its ABCI supercomputer, which provides over 37 PetaFLOPS of FP64 performance, making it the fastest AI supercomputer in Japan. It's the first ExaFLOPS- AI supercomputer in the world.
    We'll be scouring the show floor for some of the latest Tesla V100-powered systems; stay tuned.
    Reply to goldstone77
  44. Compute accelerator it is pretty much dominated by nvidia since fermi. But the most surprusing thing is in the past we thought that this market is a duopoly market between nvidia and AMD. Intel actually more successful than AMD in compute accelerator market with their Xeon Phi. Market share wise AMD is almost non existent.
    Reply to renz496
  45. DGX SaturnV Volta - NVIDIA DGX-1 Volta36, Xeon E5-2698v4 20C 2.2GHz, Infiniband EDR, NVIDIA Tesla V100 Specs:

    https://www.nvidia.com/en-us/data-center/dgx-1/

    Will the Volta GPU Help NVIDIA Conquer the Cloud? Looks Like Nvidia has it made.
    I do think that NVIDIA's latest chip (Volta) could help consolidate its lead in the fast-growing market for server GPU accelerators.

    Harsh Chauhan Pointed out why will Nvidia lead the way Nov 13, 2017 at 9:39PM
    NVIDIA's (NASDAQ:NVDA) data center business has been one of its biggest growth drivers in recent quarters thanks to the growing application of GPUs (graphics processing units) in cloud computing applications. The chipmaker now gets almost 19% of its revenue by selling GPUs for data centers, up from 10% just a year ago as almost all the big cloud players are lining up for its chips to speed up their infrastructure and prepare for the era of artificial intelligence (AI).

    In fact, NVIDIA's data center revenue shot up 175% year over year during the second quarter. More importantly, it is capable of maintaining this terrific pace of growth on the back of its recently launched Volta GPUs that are expected to enhance its lead in this space

    Amazon Web Services (AWS) recently launched its P3 cloud computing instance (an instance is a virtual server) powered by the latest Volta GPUs to target intensive applications such as autonomous vehicles, machine learning, and seismic analysis, among others. More specifically, AWS' P3 instance allows users to deploy up to eight Volta GPUs, boosting performance over the previous-generation P2 instance by as much as 14 times.

    This enhanced performance will help users train artificial intelligence models in just hours instead of days, which isn't surprising as NVIDIA claims that the new Volta chips are 12 times faster than the preceding Pascal chips. The graphics specialist has managed to eke out such terrific performance from Volta with the help of tensor cores that are meant to substantially accelerate the training and inferencing of artificial intelligence algorithms and applications.

    There has been a lot of interest in these new GPUs given the performance leap they deliver over the previous-generation architecture. The likes of HPE, IBM, Dell EMC, Huawei, and Lenovo have already announced that they will soon start using the Volta GPUs. Additionally, Chinese large-scale cloud computing providers Tencent, Baidu, and Alibaba have announced their commitment to NVIDIA's new chip, stating that they will upgrade from Pascal to Volta GPUs across data centers and related cloud infrastructure.
    Reply to YoAndy
  46. If Nvidia's Volta GPU is 132% faster than Pascal, imagine what it can do for Gamers.
    Reply to YoAndy
  47. Return to shareholders:
    NVIDIA has committed to return nearly ~$1.3 billion to shareholders in fiscal 2018. It has already returned $924 million to shareholders in fiscal 1H18. In fiscal 2H18, it’s likely to spend over $160 million on dividend payments and $160 million on share buybacks.

    NVIDIA’s capital spending strategy bodes well for its investors likely because they’re not interested in dividends, but rather in returns through stock price appreciation.

    Fiscal 2Q18 EPS
    NVIDIA’s non-GAAP (generally accepted accounting principles) EPS rose 91% YoY (year-over-year), or 19% sequentially, to $1.01 in fiscal 2Q18 because its operating profits rose faster than revenues in dollar terms. Its operating profits rose because the quarter included the sales of its very high-end GTX 1080 Ti and Titan XP, which generate higher margins.

    Intel’s (INTC) and AMD’s EPS fell 12% and 50% sequentially, respectively, during the same quarter.

    Fiscal 3Q18 EPS guidance
    Analysts expect NVIDIA to report EPS of $0.94 in fiscal 3Q18—a 7% fall from fiscal 2Q18. Fiscal 3Q is a seasonally strong quarter for NVIDIA, and it usually reports strong double-digit EPS growth during the quarter.

    The company has beaten analysts EPS estimate by an average of 36% for the past eight quarters. If the company maintains this momentum, it’s likely to report EPS of $1.28, representing 27% sequential growth.

    Long-term EPS forecast
    Analysts expect NVIDIA’s EPS to rise 17% YoY to $3.6 in fiscal 2018. The company has already earned EPS of $1.86 in fiscal 1H18, but there’s a possibility that an analyst might revise his or her EPS estimate for fiscal 2018 if the company reports EPS of over $1.0 in its fiscal 3Q18 earnings.

    For the long-term, RBC Capital Markets analyst Mitch Steves expects NVIDIA’s EPS to grow to $6.0 in fiscal 2019 and $10.0 in fiscal 2021. Evercore ISI analyst C.J. Muse expects NVIDIA’s EPS to reach $10 in the next three to five years. Both the analysts are bullish on NVIDIA.
    There is no stopping NVIDIA.
    Reply to YoAndy
  48. Take non-GAAPs analysis with a huge grain of salt though. That is the "accounting cooking book" to show good numbers.

    Still, no one can deny that nVidia is in an amazing place currently. I hope Volta keeps the good trend.

    Cheers!
    Reply to Yuka
Ask a new question Answer

Read More

Nvidia Graphics Cards Graphics