tesla hp datasheet

Download Tesla Hp Datasheet

Post on 22-Apr-2015

22 views

Category:

Documents

0 download

Embed Size (px)

TRANSCRIPT

QuickSpecsOverview

NVIDIA Tesla GPU Computing Modules for HP ProLiant Servers

HP supports, on select HP ProLiant servers, computational accelerator modules based on NVIDIA Tesla Graphical Processing Unit (GPU) technology. The following Tesla Computing Modules are available from HP, for use in HP ProLiant SL390s servers. NVIDIA Tesla M2050 2-Slot Passive Module NVIDIA Tesla M2070 2-Slot Passive Module NVIDIA Tesla M2070Q 2-Slot Passive Module NVIDIA Tesla M2090 2-Slot Passive Module The NVIDIA Tesla M2070Q module can also be used in HP ProLiant WS460c workstation blades. Based on NVIDIA's CUDA architecture, the Tesla M2050, M2070/M2070Q and M2090 Computing Modules enable seamless integration of GPU computing with HP ProLiant servers for high-performance computing and large data center, scale-out deployments. The 20-series Tesla GPUs are the first to have a peak performance greater than 10X the double-precision horsepower of a quad-core x86 CPU and the first to deliver ECC memory. The Tesla M2050, M2070/M2070Q and M2090 modules deliver all of the standard benefits of GPU computing while enabling maximum reliability and tight integration with system monitoring and management tools such as HP Cluster Management Utility. The Tesla M2070Q uses the NVIDIA Fermi GPU that combines Tesla's high performance computing - found in the Tesla M2050 and Tesla M2070 cards - and the NVIDIA Quadro professional-class visualization in the same GPU. The Tesla M2070Q is the ideal solution for customers who want to deploy high performance computing in addition to advanced and remote visualization in the same datacenter. The HP GPU Ecosystem includes HP Cluster Platform specification and qualification, HP-supported GPU-aware cluster software, and also third-party GPU-aware cluster software for NVIDIA Tesla GPU Computing Modules on HP ProLiant Servers. In particular, the HP Cluster Management Utility (CMU) will monitor and display GPU health sensors such as temperature. CMU will also install and provision the GPU drivers and the CUDA software. The HP HPC Linux Value Pack includes a GPU-enhanced version of Platform LSF, with the capability of scheduling jobs based on GPU requirements. The HP HPC Linux Value pack also includes a GPU-enhanced version of HP-MPI, which can set up optimized affinities between specific cores and specific GPUs.

DA - 13743

Worldwide Version 6 June 10, 2011

Page 1

QuickSpecsOverview

NVIDIA Tesla GPU Computing Modules for HP ProLiant Servers

What's NewSupport for NVIDIA M2090 Tesla GPU Module

DA - 13743

Worldwide Version 6 June 10, 2011

Page 2

QuickSpecsModelsNVIDIA Passive Tesla Modules

NVIDIA Tesla GPU Computing Modules for HP ProLiant Servers

NVIDIA Tesla M2050 3GB Module NOTE: 2-slot passively cooled second-generation Tesla module with 3 GB memory. NVIDIA Tesla M2070Q 6GB GPU Graphics Module NVIDIA Tesla M2070 6GB Module NVIDIA Tesla M2090 6GB GPU Module NOTE: 2-slot passively cooled second-generation Tesla modules with 6 GB memory. NOTE: See the HP ProLiant SL390s Generation 7 (G7) server or HP ProLiant WS460c (G6) Workstation Blade QuickSpecs for configuration details: http://h18004.www1.hp.com/products/quickspecs/13713_div/13713_div.html http://h18004.www1.hp.com/products/quickspecs/13429_div/13429_div.html

SH885B A0C39A SH886A A0J99A

DA - 13743

Worldwide Version 6 June 10, 2011

Page 3

QuickSpecsStandard Features

NVIDIA Tesla GPU Computing Modules for HP ProLiant Servers

M2050, M2070/M2070Q and M2090 Computing ModulesPerformance of M2050 and M2070/M2070Q Computing Modules 448 CUDA cores 515 Gigaflops of double-precision peak performance in each GPU Single precision peak performance is over one Teraflop per GPU. GDDR5 memory optimizes performance and reduces data transfers by keeping large data sets in local memory (6 GB on the M2070 module) that is attached directly to the GPU. The NVIDIA Parallel DataCache accelerates algorithms such as physics solvers, ray-tracing, and sparse matrix multiplication where data addresses are not known beforehand. This includes a configurable L1 cache per Streaming Multiprocessor block and a unified L2 cache for all of the processor cores. The NVIDIA GigaThread Engine maximizes the throughput by faster context switching that is 10X faster than the M1060 module, concurrent kernel execution, and improved thread block scheduling. Asynchronous transfer turbo charges system performance by transferring data over the PCIe bus while the computing cores are crunching other data. Even applications with heavy data-transfer requirements, such as seismic processing, can maximize the computing efficiency by transferring data to local memory before it is needed. The high speed PCIe Gen 2.0 data transfer maximizes bandwidth between the HP ProLiant server and the Tesla processors. Performance of the M2090 Computing Module 512 CUDA cores 655 Gigaflops of double-precision peak performance in each GPU 1330 Gigaflops of single-precision peak performance in each GPU. GDDR5 memory optimizes performance and reduces data transfers by keeping large data sets in 6 GB of local memory that is attached directly to the GPU. The NVIDIA Parallel DataCache accelerates algorithms such as physics solvers, ray-tracing, and sparse matrix multiplication where data addresses are not known beforehand. This includes a configurable L1 cache per Streaming Multiprocessor block and a unified L2 cache for all of the processor cores. The NVIDIA GigaThread Engine maximizes the throughput by faster context switching that is 10X faster than the M1060 module, concurrent kernel execution, and improved thread block scheduling. Asynchronous transfer turbo charges system performance by transferring data over the PCIe bus while the computing cores are crunching other data. Even applications with heavy data-transfer requirements, such as seismic processing, can maximize the computing efficiency by transferring data to local memory before it is needed. The high speed PCIe Gen 2.0 data transfer maximizes bandwidth between the HP ProLiant server and the Tesla processors. Reliability ECC Memory meets a critical requirement for computing accuracy and reliability for datacenters and supercomputing centers. It offers protection of data in memory to enhance data integrity and reliability for applications. Register files, L1/L2 caches, shared memory, and DRAM all are ECC protected. Double-bit errors are detected and can trigger alerts with the HP Cluster Management Utility. Also, the Platform LSF job scheduler, available as part of HP HPC Linux Value Pack, can be configured to report when jobs encounter double-bit errors. Passive heatsink design eliminates moving parts and cables reduces mean time between failures. Programming and Management Ecosystem The CUDA programming environment has broad support of programming languages and APIs. Choose C, C++, OpenCL, DirectCompute, or Fortran to express application parallelism and take advantage of the innovative "Fermi" architecture. The CUDA software, as well as the GPU drivers, can be automatically installed on HP ProLiant servers, by HP Cluster Management Utility. "Exclusive mode" enables application-exclusive access to a particular GPU. CUDA environment variables enable cluster

DA - 13743

Worldwide Version 6 June 10, 2011

Page 4

QuickSpecsStandard Features

NVIDIA Tesla GPU Computing Modules for HP ProLiant Servers

management software such as the Platform LSF job scheduler (available as part of HP HPC Linux Value Pack) to limit the Tesla GPUs an application can use. With HP ProLiant servers, application programmers can control the mapping between processes running on individual cores, and the GPUs with which those processes communicate. By judicious mappings, the GPU bandwidth, and thus overall performance, can be optimized. The technique is described in a white paper available to HP customers at: www.hp.com/go/hpc. A heuristic version of this affinity-mapping has also been implemented by HP as an option to the mpirun command as used for example with HP-MPI, available as part of HP HPC Linux Value Pack. GPU control is available through the nvidia-smi tool which lets you control compute-mode (e.g. exclusive), enable/disable/report ECC and check/reset double-bit error count. IPMI and iLO gather data such as GPU temperature. HP Cluster Management Utility has incorporated these sensors into its monitoring features so that cluster-wide GPU data can be presented in real time, can be stored for historical analysis and can be easily used to set up management alerts. Supported Operating Systems RHEL 5 RHEL 6 SLES 11 Windows Server 2008 HP ProLiant SL390s G7 (M2050, M2070/M2070Q,M2090) NOTE: The ambient temperature for SL390s 2U systems with between one and three NVIDIA M2090 GPUs, must be 30 degrees Celsius or less. The ambient temperature for SL390s 4U systems with between five and eight NVIDIA M2090 GPUs, must be 30 degrees Celsius or less. All other SL390s systems may be operated with ambient temperatures up to 35 degrees Celsius. HP ProLiant WS460c G6 (M2070Q only)

Supported Servers and Workstation Blades

HP Services and Support The NVIDIA Tesla GPU Computing Module has one year for parts exchange only or the warranty of the server or chassis it is attached to and for which it is qualified. Enhancements to warranty services are available for server and chassis through Flexible Care Pack services. NOTE: For more information, visit HP Care Pack Services at: http://www.hp.com/services

DA - 13743

Worldwide Version 6 June 10, 2011

Page 5

QuickSpecsOptional FeaturesHP High Performance Clusters HP Cluster Platforms

NVIDIA Tesla GPU Computing Modules for HP ProLiant Servers

The NVIDIA Tesla GPU Computing Modules are opti

Recommended

View more >