NVIDIA Announces “NVIDIA Titan V" Video Card: GV100 for $3000, On Sale Now
by Ryan Smith & Nate Oh on December 7, 2017 11:45 PM EST- Posted in
- GPUs
- NVIDIA
- Volta
- Neural Networks
- GV100
- Tensor
- Deep Learning
- Titan V
Out of nowhere, NVIDIA has revealed the NVIDIA Titan V today at the 2017 Neural Information Processing Systems conference, with CEO Jen-Hsun Huang flashing out the card on stage. A mere 7 months after Volta was announced with the Tesla V100 accelerator and the GV100 GPU inside it, NVIDIA continues its breakneck pace by releasing the GV100-powered Titan V, available for sale today. Aimed at a decidedly more compute-oriented market than ever before, the 815 mm2 behemoth die that is GV100 is now available to the broader public.
NVIDIA Compute Accelerator Specification Comparison | ||||||
Titan V | Tesla V100 (PCIe) |
Tesla P100 (PCIe) |
Titan Xp | |||
CUDA Cores | 5120 | 5120 | 3584 | 3840 | ||
Tensor Cores | 640 | 640 | N/A | N/A | ||
Core Clock | 1200MHz | ? | ? | 1485MHz | ||
Boost Clock | 1455MHz | 1370MHz | 1300MHz | 1582MHz | ||
Memory Clock | 1.7Gbps HBM2 | 1.75Gbps HBM2 | 1.4Gbps HBM2 | 11.4Gbps GDDR5X | ||
Memory Bus Width | 3072-bit | 4096-bit | 4096-bit | 384-bit | ||
Memory Bandwidth | 653GB/sec | 900GB/sec | 720GB/sec | 547GB/sec | ||
VRAM | 12GB | 16GB | 16GB | 12GB | ||
L2 Cache | 4.5MB | 6MB | 4MB | 3MB | ||
Single Precision | 13.8 TFLOPS | 14 TFLOPS | 9.3 TFLOPS | 12.1 TFLOPS | ||
Double Precision | 6.9 TFLOPS (1/2 rate) |
7 TFLOPS (1/2 rate) |
4.7 TFLOPS (1/2 rate) |
0.38 TFLOPS (1/32 rate) |
||
Tensor Performance (Deep Learning) |
110 TFLOPS | 112 TFLOPS | N/A | N/A | ||
GPU | GV100 (815mm2) |
GV100 (815mm2) |
GP100 (610mm2) |
GP102 (471mm2) |
||
Transistor Count | 21.1B | 21.1B | 15.3B | 12B | ||
TDP | 250W | 250W | 250W | 250W | ||
Form Factor | PCIe | PCIe | PCIe | PCIe | ||
Cooling | Active | Passive | Passive | Active | ||
Manufacturing Process | TSMC 12nm FFN | TSMC 12nm FFN | TSMC 16nm FinFET | TSMC 16nm FinFET | ||
Architecture | Volta | Volta | Pascal | Pascal | ||
Launch Date | 12/07/2017 | Q3'17 | Q4'16 | 04/07/2017 | ||
Price | $2999 | ~$10000 | ~$6000 | $1299 |
For the spec sheet we've gone ahead and lined it up against NVIDA's other Pascal cards, and for good reason. While the Titan series of cards may have started life as a prosumer card in 2013, since then NVIDIA's GPU designs have become increasingly divergent between compute and graphics. And even though the previous Titan Xp was based on the more graphics-focused GP102 GPU, the card itself was primarily (but not solely) pitched as an entry-level compute card, for customers who needed a (relatively) cheap way to do FP32 compute and neural network inferencing in workstations and small clusters.
The Titan V, by extension, sees the Titan lineup finally switch loyalties and start using NVIDIA’s high-end compute-focused GPUs, in this case the Volta architecture based V100. The end result is that rather than being NVIDIA’s top prosumer card, the Titan V is decidedly more focused on compute, particularly due to the combination of the price tag and the unique feature set that comes from using the GV100 GPU. Which isn’t to say that you can’t do graphics on the card – this is still very much a video card, outputs and all – but NVIDIA is first and foremost promoting it as a workstation-level AI compute card, and by extension focusing on the GV100 GPU’s unique tensor cores and the massive neural networking performance advantages they offer over earlier NVIDIA cards.
In this sense the Titan V is a return to form of sorts to the professional side of prosumer for the Titan family. One of the original claims to fame for the original Titan was its high performance in specialized FP64 compute workloads, something that was lost on the later Titan X and Titan Xp. By switching to NVIDIA’s specialized high-end compute GPUs, the Titan V regains its formerly lost compute capabilities, all the while also gaining all of the compute capabilities NVIDIA has introduced since then. It’s no mistake that Jen-Hsun introduced the card at a neural networking conference, as this is a big chunk of the professional computing audience that NVIDIA is targeting with the card.
Interestingly, comparing it to the PCIe Tesla V100, I’m surprised by just how close the cards are in features and performance. NVIDIA has confirmed that the Titan V gets the GV100 GPU’s full, unrestricted FP64 compute and tensor core performance. To the best of our knowledge (and from what NVIDIA will comment on) it doesn’t appear that they’ve artificially disabled any of the GPU’s core features. What does separate the Titan from the Tesla then from a performance standpoint is quite simple: memory capacity, memory bandwidth, and the lack of NVLink functionality. There are also a number of smaller differences between the cards that help to differentiate them between server and workstation – such as passive versus active cooling, NVLink, and the support policies – but otherwise for customers who are running a small number of cards, the Titan V’s feature set is remarkably close to the much more expensive Tesla V100’s, which is a very interesting development since it goes to show just how confident NVIDIA is that this won’t undermine Tesla sales.
Moving on and diving into the numbers, Titan V features 80 streaming multiprocessors (SMs) and 5120 CUDA cores, the same amount as its Tesla V100 siblings. The differences come with the memory and ROPs. In what's clearly a salvage part for NVIDIA, one of the card's 4 memory partitions has been cut, leaving Titan V with 12GB of HBM2 attached via a 3072-bit memory bus. As each memory controller is associated with a ROP partition and 768 KB of L2 cache, this in turn brings L2 down to 4.5 MB, as well as cutting down the ROP count.
In terms of clockspeeds, the HBM2 has been downclocked slightly to 1.7GHz, while the 1455MHz boost clock actually matches the 300W SXM2 variant of the Tesla V100, though that accelerator is passively cooled. Notably, the number of tensor cores have not been touched, though the official 110 DL TFLOPS rating is lower than the 1370MHz PCIe Tesla V100, as it would appear that NVIDIA is using a clockspeed lower than their boost clock in these calculations.
For the card itself, it features a vapor chamber cooler with copper heatsink and 16 power phases, all for the 250W TDP that has become standard with the single GPU Titan models. Output-wise, the Titan V brings 3 DisplayPorts and 1 HDMI connector. And as for card-to-card communication, PCB itself appears to have NVLink connections on the top, but these look to have been intentionally blocked by the shroud to prevent their use and are presumably disabled.
As mentioned earlier, NVIDIA is unsurprisingly pushing this as a compute accelerator card, especially considering that Titan V features tensor cores and keeps the TITAN branding as opposed to GeForce TITAN. But there are those of us who know better than to assume people won’t drop $3000 to use the latest Titan card for gaming, and while gaming is not the primary (or even secondary) focus of the card, you also won't see NVIDIA denying it. In that sense the Titan V is going to be treated as a jack-of-all-trades card by the company.
To that end, no gaming performance information has been disclosed, but NVIDIA has confirmed that the card uses the standard GeForce driver stack. And on that note, yesterday NVIDIA released 388.59 bringing official Titan V support. Now, how much those drivers have actually been optimized for the GV100 is another matter entirely; Volta is a new architecture, markedly so at times. Speaklng solely off the cuff here, for graphics workloads the card has more resources than the Titan Xp in almost every meaningful metric, but it's also a smaller difference on paper than you might think.
As for NVIDIA's intended market of compute and AI users, the Titan V will be supported by NVIDIA GPU Cloud, which includes TensorRT, a number of deep learning frameworks, and HPC-related tools.
If the golden shroud didn’t already suggest so, the Titan V is also carving out a new eye-watering price point, dropping in at $2999 and on sale now at the NVIDIA store. NVIDIA has, to date, been selling Tesla V100 products as fast as they can produce them, so I'm not going to be surprised if the Titan V sees a similar fate. The $3000 price tag is quite high, even by Titan standards, but with the rare Tesla V100 PCIe card going for around $10,000, the Titan V is markedly cheaper. In fact in some respects I'm surprised NVIDIA is selling a GV100 card for so little; these are GV100 salvage parts that don't make the cut for Tesla - so the alternative would be throwing them away - but it just goes to show how confident NVIDIA is that it won't undermine the Tesla family.
At any rate, for NVIDIA professional users who have been looking to dip their toes into Volta but didn't want a full-fledged Tesla card, the Titan V is clearly going to be a popular card. Over the last two years NVIDIA's AI efforts have been firing on all cylinders, and by bringing a GV100 card down to just $3000, expect to see them crack open the market that much further. I dare say the idea of the "prosumer" Titan has died with this card, but for the rapidly growing professional compute market, this looks to be exactly the kind of card that a lot of developers have been waiting for.
Update (12/8/17): Yesterday, NVIDIA also released driver version 388.59 WHQL, bringing product support to the Titan V, along with Fallout 4 VR support. NVIDIA has noted that the Titan V currently suffers from TDR errors and display blanking during Blu-ray disc playback on high resolutions, as well as from G-Sync display blanking when repeatedly switching between different memory overclocks. Lastly, this minor update features one bug fix, resolving flickering on GTX 1080 equipped G-Sync notebooks.
159 Comments
View All Comments
Holliday75 - Friday, December 8, 2017 - link
We all know its dead.HighTech4US - Friday, December 8, 2017 - link
Yes it can.The GTX 1080 Ti can play Crysis @ 3840x2160 Very High settings:
Even fewer of Nvidia’s cards cooperated with our 4K screen, so we pulled the plug at GeForce GTX 780 Ti. It was for the best, perhaps, as that 2013-era card averaged just over 30 FPS.
The real news, however, is GeForce GTX 1080 Ti, averaging nearly 80 FPS at 4K in Crysis. Ten years on, and there’s exactly one gaming graphics card (we’re not counting Titan cards here) able to do this. Everything else from AMD and Nvidia fails to break the symbolic 60 FPS mark.
http://www.tomshardware.com/reviews/crysis-10-year...
Since the Titan V is more powerful than the 1080 Ti it would do even better.
Gc - Saturday, December 9, 2017 - link
But can it play Crysis as well as a young human?If so, are gamers obsolete? /s
(Or choose your own autonomous task.)
Yojimbo - Friday, December 8, 2017 - link
This is a cheaper V100 more than a faster 1080 Ti. It's not a GeForce card. It's a true prosumer card. It's got nothing to do with AMD. The card is only out because NVIDIA think there is a market for it at $3,000. As long as they can charge $8000+ for the V100, there isn't a world where they sell the Titan V for $1,000 or $1,200. They'd be compromising their own data center line.Of course NVIDIA will happily sell the card to anyone who wants the absolute fastest card for gaming they can get right now. It probably won't be surpassed until the 1180 Ti (or 2080 Ti or whatever they call it) comes out in a year or so. But that's a pretty rich premium to pay for a year of the fastest gaming card on Earth.
beginner99 - Friday, December 8, 2017 - link
Exactly. At work we are thinking in investing into deep learning but cost of tesla card that are much weaker than this one are easily $6000++. So $3000 is a steal for deep learning application assuming there is no undisclosed limitation in speed.WB312 - Friday, December 8, 2017 - link
Oh boy Volta GPUs are gonna be expensive with no competition from AMD in the high end.silverblue - Friday, December 8, 2017 - link
They're going to be expensive anyway. NVIDIA did continue to sell Titan X at a stupid price when the 295 X2 was creaming it for half the price, if memory serves.Klimax - Friday, December 8, 2017 - link
If you were lucky to have Crossfire to work. (SLI was bit better, but it wasn't "universal" solution either)Heh, AMD needed two chips to beat one...
Manch - Friday, December 8, 2017 - link
I think the 2 series was when AMD abandoned its little core/multi core idea and went back to making big cores. Yup still didn't work. AMDs HW isnt bad but their drivers fail time and time again to actually utilize the GPUs effectively.Neither Crossfire or SLI are particularly great. When they work, yeah, but more often than not one card is just sitting there. I'll probably take my second 290 and put it in another rig so the wife can have her own DT and then I'll get her an oculus so she can fill up her SSD with crap instead of mine :'D
silverblue - Friday, December 8, 2017 - link
I think AMD have gone from having drivers that were often buggy and not able to expose the full potential of their cards, to having drivers that are more stable... and still not able to expose the full potential of their cards. Is Vega's Draw Stream Binning Rasterizer supported yet?