from Hacker News

Nvidia Announces A100 80GB GPU for AI

by alphadevx on 11/18/20, 8:49 PM with 154 comments

  • by fxtentacle on 11/18/20, 10:04 PM

    While these are amazing performance numbers, I seriously wonder if all this expensive AI magic will be cost effective in the end.

    Their lead example is recommendation systems, but I can't say that I have received many good suggestions recently.

    Spotify and Deezer both suggest the chart hits, regardless of how often I dislike that kind of music.

    Amazon keeps recommending me tampons (I'm a guy) ever since I've had a coworking office with a female colleague in 2015.

    For all the data that they collect and all the AI that they pay for, these companies get very little revenue to show for it.

  • by john_moscow on 11/18/20, 9:46 PM

    I am more wondering why hasn't AMD massively invested into porting common ML frameworks to OpenCL. Nvidia has outrageous margins on their datacenter GPUs. They've even banned the use of lower-margin gamer-oriented GPU in datacenters [0]. Given that tensor arithmetic is essentially an easily abstractable commodity, I just don't understand why they don't offer a drop-in replacement.

    Most users won't care what hardware their PyTorch model runs on in the cloud. All that matters for them is dollars per training epoch (or cents per inference). This could be a steal for an alternate hardware vendor.

    [0] https://web.archive.org/web/20201109023551/https://www.digit...

  • by ogre_codes on 11/18/20, 11:25 PM

    It occurs to me that if anyone is going to release an ARM based CPU that is competitive with Apple, it's Nvidia. A Microsoft/ Nvidia partnership could create some pretty impressive Surface laptops, and if Nvidia were allowed to sell those CPUs to OEMs for other 2 in 1s or laptops, Microsoft might just get some traction on their ARM efforts.
  • by hn3333 on 11/18/20, 9:27 PM

    Does processing power translate to actual power? Will most stock market gains in the end go to the people with the fastest computers? Will wars be won by the groups with the most processing speed? Was Cyberpunk (slightly) wrong and it's just about having more memory and more instructions per millisecond than the rest? Are sophisticated circuits the new oil?
  • by peter_d_sherman on 11/19/20, 7:44 AM

    >"The new A100 with HBM2e technology doubles the A100 40GB GPU’s high-bandwidth memory to 80GB and delivers

    over 2 terabytes per second of memory bandwidth."

  • by eigenvalue on 11/19/20, 2:44 AM

    The workstation they announced in the press release [0] sounds and looks incredible. I'm sure it costs over $100k, but I wish that kind of case design and cooling system could be available at a more reasonable price point. I wonder how long it will take for a computer with the performance specs of the DGX Station A100 to be available for under $3,000. Will that take 5 years? 10 years? Based on the historical trends of the last decade, those estimates strike me as pretty optimistic.

    [0] https://www.nvidia.com/en-us/data-center/dgx-station-a100/

  • by vbezhenar on 11/18/20, 9:46 PM

    Why do they still produce GPU rather than specialized ASIC for neural networks like Google does with their Tensor Processing Units?
  • by phs318u on 11/19/20, 2:19 AM

    Legit question: Why are these still called "GPU"s? Shouldn't they rightly be called "AIPU"s, or "IPU"s?
  • by sien on 11/19/20, 6:38 AM

    It is amusing to see this so soon after a post on how workstations were dead : https://news.ycombinator.com/item?id=24977652

    $US 200K for startling performance this time.

  • by taf2 on 11/19/20, 12:56 AM

    Any word on how expensive this board will be?
  • by Rafuino on 11/18/20, 10:23 PM

    For only the low low cost of everything in your bank account!