Nvidia Tesla

This article is about GPGPU cards. For the GPU microarchitecture, see Tesla (microarchitecture).
Nvidia Tesla

Nvidia Tesla is Nvidia's brand name for their products targeting stream processing and/or general purpose GPU. Products use GPUs from the G80 series onward. Both the underlying microarchitecture of the initial GPUs "Tesla" and the Tesla product line take their name from pioneering electrical engineer Nikola Tesla.

Overview

With their very high computational power (measured in floating point operations per second or FLOPS) compared to microprocessors, the Tesla products target the high performance computing market.[1] As of 2012, Nvidia Teslas power some of the world's fastest supercomputers, including Titan at Oak Ridge National Laboratory and Tianhe-1A, in Tianjin, China.

The lack of ability to output images to a display was the main difference between Tesla products and the consumer level GeForce cards and the professional level Quadro cards, but the latest Tesla C-class products include one Dual-Link DVI port.[2] For equivalent single precision output, Fermi-based Nvidia GeForce cards have four times less dual-precision performance. Tesla products primarily operate:[3]

Nvidia intends to offer ARMv8 processor cores embedded into future Tesla GPUs as part of Project Denver.[4] This will be a 64-bit follow on to the 32-bit Tegra chips.

Market

The defense industry currently accounts for less than a sixth of Tesla sales, but Sumit Gupta predicts further sales to the geospatial intelligence market.[5]

Specifications and configurations

Nvidia Tesla C2075


Model Micro-
architecture
Chips Core clock
(MHz)
Shaders Memory Processing power (GFLOPS)5 CUDA
compute
ability8
TDP
(watts)
Notes, form_factor
Thread processors
(total)
Base clock (MHz) Max Boost
Clock (MHz)7
Bus type Bus width
(bit)
Size
(MB)
Clock
(MT/s)
Bandwidth
(GB/s)
Single precision
(MAD+MUL)
Single precision
(MAD or FMA)
Double precision
(FMA)
Units MHz MHz W
C870 GPU Computing Module1 Tesla 1× G80 600 128 1350 N/A GDDR3 384 1536 1600 76.8 518.4 345.6 No 1.0 170.9 Internal PCIe GPU (full-height, dual-slot)
D870 Deskside Computer1 2× G80 600 256 1350 N/A GDDR3 2× 384 2× 1536 1600 2× 76.8 1036.8 691.2 No 1.0 520 Deskside or 3U rack-mount external GPUs
S870 GPU Computing Server1 4× G80 600 512 1350 N/A GDDR3 4× 384 4× 1536 1600 4× 76.8 2073.6 1382.4 No 1.0 1U rack-mount external GPUs, connect via 2× PCIe (×16)
C1060 GPU Computing Module2 1× GT200 602 240 1296[7] N/A GDDR3 512 4096 1600 102.4 933.12 622.08 77.76 1.3 187.8 Internal PCIe GPU (full-height, dual-slot)
S1070 GPU Computing Server "400 configuration"2 4× GT200 602 960 1296 N/A GDDR3 4× 512 4× 4096 1538.4 4× 98.5 3732.5 2488.3 311.0 1.3 800 1U rack-mount external GPUs, connect via 2× PCIe (×8 or ×16)
S1070 GPU Computing Server "500 configuration"2 1440 N/A 4147.2 2764.8 345.6
S1075 GPU Computing Server2[8] 4× GT200 602 960 1440 N/A GDDR3 4× 512 4× 4096 1538.4 4× 98.5 4147.2 2764.8 345.6 1.3 1U rack-mount external GPUs, connect via 1× PCIe (×8 or ×16)
Quadro Plex 2200 D2 Visual Computing System6 2× GT200GL 648 480 1296 N/A GDDR3 2× 512 2× 4096 1600 2× 102.4 1866.2 1244.2 155.5 1.3 Deskside or 3U rack-mount external GPUs with 4 dual-link DVI outputs
Quadro Plex 2200 S4 Visual Computing System6 4× GT200GL 648 960 1296 N/A GDDR3 4× 512 4× 4096 1600 4× 102.4 3732.5 2488.3 311.0 1.3 1200 1U rack-mount external GPUs, connect via 2× PCIe (×8 or ×16)
C2050 GPU Computing Module[9] Fermi 1× GF100 575 448 1150 N/A GDDR5 384 30724 3000 144 No 1030.4 515.2 2.0 247 Internal PCIe GPU (full-height, dual-slot)
M2050 GPU Computing Module[10] N/A 3092 148.4 No 225
C2070 GPU Computing Module[9] 1× GF100 575 448 1150 N/A GDDR5 384 61444 3000 144 No 1030.4 515.2 2.0 247 Internal PCIe GPU (full-height, dual-slot)
C2075 GPU Computing Module[11] N/A 3000 144 No 225
M2070/M2070Q GPU Computing Module[12] N/A 3132 150.336 No 225
M2090 GPU Computing Module[13] 1× GF110 650 512 1300 N/A GDDR5 384 61444 3700 177.6 No 1331.2 665.6 2.0 225 Internal PCIe GPU (full-height, dual-slot)
S2050 GPU Computing Server 4× GF100 575 1792 1150 N/A GDDR5 4× 384 4× 30724 3092 4× 148.4 No 4121.6 2060.8 2.0 900 1U rack-mount external GPUs, connect via 2× PCIe (×8 or ×16)
S2070 GPU Computing Server N/A 4× 61444 No
K10 GPU Accelerator[14] Kepler 2× GK104 N/A 3072 745 ? GDDR5 2× 256 2× 4096 5000 2× 160 No 4577 190.7 3.0 225 Internal PCIe GPU (full-height, dual-slot)
K20 GPU Accelerator[15][16] 1× GK110 N/A 2496 706 ? GDDR5 320 5120 5200 208 No 3524 1175 3.5 225 Internal PCIe GPU (full-height, dual-slot)
K20X GPU Accelerator[17] 1× GK110 N/A 2688 732 ? GDDR5 384 6144 5200 250 No 3935 1312 3.5 235 Internal PCIe GPU (full-height, dual-slot)
K40 GPU Accelerator[18] 1× GK110B N/A 2880 745 875 GDDR5 384 122884 6000 288 No 4291–5040 1430–1680 3.5 235 Internal PCIe GPU (full-height, dual-slot)
K80 GPU Accelerator[19] 2× GK210 N/A 4992 560 875 GDDR5 2× 384 2× 12288 5000 2× 240 No 5591–8736 1864–2912 3.7 300 Internal PCIe GPU (full-height, dual-slot)
M10 GPU Accelerator[20] Maxwell 4x GM107 N/A 4x 640 405 1033 GDDR5 4x 128 4x 8192 4 x 83 No 4x 518-1322 5.2 225 Internal PCIe GPU (full-height, dual-slot)
M4 GPU Accelerator[21][22] 1× GM206 N/A 1024 872 1072 GDDR5 128 4096 5500 88 No 1786–2195 55.81–68.61 5.2 50–75 Internal PCIe GPU (half-height, single-slot)
M40 GPU Accelerator[22][23] 1× GM200 N/A 3072 948 1114 GDDR5 384 12288 6000 288 No 5825–6844 182.0–213.9 5.2 250 Internal PCIe GPU (full-height, dual-slot)
M6 GPU Accelerator[24] 1× GM204 N/A 1536 722 1051 GDDR5 256 8192 4600 147.2 No 2218–3229 69.3–100.9 5.2 75–100 Internal MXM GPU
M60 GPU Accelerator[25] 2× GM204 N/A 4096 899 1178 GDDR5 2× 256 2× 8192 5000 2× 160 No 7365–9650 230.1–301.6 5.2 225–300 Internal PCIe GPU (full-height, dual-slot)
P4 GPU Accelerator[26] Pascal 1× GP104 N/A 2560 810 1063 GDDR5 256 8192 6000 192.0 No 4147–5443 129.6–170.1 6.1 50-75 PCIe card
P40 GPU Accelerator[26] 1× GP102 N/A 3840 1303 1531 GDDR5 384 24576 7200 345.6 No 10007–11758 312.7–367.4 6.1 250 PCIe card
P100 GPU Accelerator (Mezzanine)[27][28] GP100 [29] N/A 3584 1328 1480 HBM2 4096 16384 1406 x8 720 No, 21200 (FP16) 9519–10609 4760–5304 6.0 300 NVLink card
P100 GPU Accelerator (16 GB Card)[30] N/A 1126 1303 No, 18700 (FP16) 8071‒9340 4036‒4670 250 PCIe card
P100 GPU Accelerator (12 GB Card)[30] N/A 3072 12288 540 No, 18700 (FP16) 8071‒9340 4036‒4670
Quadro DGX-1,[31][32] 8x GP100 N/A 3584 x8 1328 1480 4096 16384 x8 720 x8 No, 170000 (FP16) 10000 x8 5000 x8 6.0 3200 NVLink
Model Micro-
architecture
Chips Core clock
(MHz)
Shaders Memory Processing power (GFLOPS)5 CUDA
compute
ability
TDP
(watts)
Notes, form factor
Thread processors
(total)
Base clock (MHz) Max Boost
Clock (MHz)7
Bus type Bus width
(bit)
Memory
(MB)
Clock
(MT/s)
Bandwidth
(total)
(GB/s)
Single precision
(MAD+MUL)
Single precision
(MAD or FMA)
Double precision
(FMA)

See also

References

  1. High Performance Computing - Supercomputing with Tesla GPUs
  2. Tesla Technical Brief (PDF)
  3. "Nvidia to Integrate ARM Processors in Tesla." (Article from 2012)
  4. "Nvidia chases defense, intelligence ISVs with GPUs."
  5. "Nvidia GPU Boost For Tesla" (PDF). January 2014. Retrieved 7 December 2015.
  6. "Tesla C1060 Computing Processor Board" (PDF). Nvidia.com. Retrieved 2015-12-11.
  7. "Difference between Tesla S1070 and S1075". 31 October 2008. Retrieved December 2015. S1075 has one interface card Check date values in: |access-date= (help)
  8. 1 2 "Tesla C2050 and Tesla C2070 Computing Processor" (PDF). Nvidia.com. Retrieved 2015-12-11.
  9. "Tesla M2050 and Tesla M2070/M2070Q Dual-Slot Computing Processor Modules" (PDF). Nvidia.com. Retrieved 2015-12-11.
  10. "Tesla C2075 Computing Processor Board" (PDF). Nvidia.com. Retrieved 2015-12-11.
  11. Hand, Randall (2010-08-23). "NVidia Tesla M2050 & M2070/M2070Q Specs OnlineVizWorld.com". VizWorld.com. Retrieved 2015-12-11.
  12. "Tesla M2090 Dual-Slot Computing Processor Module" (PDF). Nvidia.com. Retrieved 2015-12-11.
  13. "Tesla K10 GPU Accelerator" (PDF). Nvidia.com. Retrieved 2015-12-11.
  14. "Tesla K20 GPU Active accelerator" (PDF). Nvidia.com. Retrieved 2015-12-11.
  15. "Tesla K20 GPU Accelerator" (PDF). Nvidia.com. Retrieved 2015-12-11.
  16. "Tesla K20X GPU Accelerator" (PDF). Nvidia.com. Retrieved 2015-12-11.
  17. "Tesla K40 GPU Accelerator" (PDF). Nvidia.com. Retrieved 2015-12-11.
  18. "Tesla K80 GPU Accelerator" (PDF). Images.nvidia.com. Retrieved 2015-12-11.
  19. "Tesla M10" (PDF). Images.nvidia.com. Retrieved 2016-10-29.
  20. "Nvidia Announces Tesla M40 & M4 Server Cards - Data Center Machine Learning". Anandtech.com. Retrieved 2015-12-11.
  21. 1 2 "Accelerating Hyperscale Datacenter Applications with Tesla GPUs | Parallel Forall". Devblogs.nvidia.com. 2015-11-10. Retrieved 2015-12-11.
  22. "Tesla M40" (PDF). Images.nvidia.com. Retrieved 2015-12-11.
  23. "Tesla M6" (PDF). Images.nvidia.com. Retrieved 2016-05-28.
  24. "Tesla M60" (PDF). Images.nvidia.com. Retrieved 2016-05-27.
  25. 1 2 Smith, Ryan (13 September 2016). "Nvidia Announces Tesla P40 & Tesla P4 - Network Inference, Big & Small". Anandtech. Retrieved 13 September 2016.
  26. Smith, Ryan (5 April 2016). "Nvidia Announces Tesla P100 Accelerator - Pascal GP100 for HPC". Anandtech.com. Anandtech.com. Retrieved 5 April 2016.
  27. Harris, Mark. "Inside Pascal: Nvidia's Newest Computing Platform". Retrieved 13 September 2016.
  28. https://images.nvidia.com/content/pdf/tesla/whitepaper/pascal-architecture-whitepaper.pdf
  29. 1 2 Smith, Ryan (20 June 2016). "NVidia Announces PCI Express Tesla P100". Anandtech.com. Retrieved 21 June 2016.
  30. http://images.nvidia.com/content/quadro/vca/pdf/167106-dgx-1-datasheet-siggraph-july16-r6.pdf
  31. http://www.nvidia.com/object/deep-learning-system.html
Wikimedia Commons has media related to Nvidia Tesla series.
This article is issued from Wikipedia - version of the 11/5/2016. The text is available under the Creative Commons Attribution/Share Alike but additional terms may apply for the media files.