Based on the next-generation CUDA™ architecture codenamed “Fermi”, the Tesla M2050 and M2070 Computing Modules enable seamless integration of GPU computing with host systems for high-performance computing and large data center, scale-out deployments. The 20-series Tesla GPUs are the first to deliver greater than 10X the double-precision horsepower of a quad-core x86 CPU and the first to deliver ECC memory. The Tesla M2050 and M2070 modules deliver all of the standard benefits of GPU computing while enabling maximum reliability and tight integration with system monitoring and management tools. This gives data center IT staff much greater choice in how they deploy GPUs, with a wide variety of rack-mount and blade systems and with the remote monitoring and remote management capabilities they need.
Compared to CPU-only systems, servers with Tesla 20-series GPU Computing Modules deliver supercomputing power at 1/10th the cost and 1/20th the power consumption while providing the highest compute density.
Find out more details from participating OEM resellers.
Features
GPUs powered by Fermi-generation CUDA architecture Delivers cluster performance at 1/10th the cost and 1/20th the power of CPU-only systems based on the latest quad-core CPUs.
448 CUDA Cores Delivers up to 515 Gigaflops of double-precision peak performance in each GPU, enabling servers from leading OEMs to deliver a Teraflop or more of double-precision performance per 1 RU of space. Single precision peak performance is over one Teraflop per GPU.
ECC Memory Meets a critical requirement for computing accuracy and reliability for datacenters and supercomputing centers. Offers protection of data in memory to enhance data integrity and reliability for applications. Register files, L1/L2 caches, shared memory, and DRAM all are ECC protected.
Up to 6GB of GDDR5 memory per GPU Maximizes performance and reduces data transfers by keeping larger data sets in local memory that is attached directly to the GPU.
System Monitoring Features Integrates the GPU subsystem with the host system's monitoring and management capabilities. This means IT staff can manage all of the critical components of the computing system through a common management interface such as IPMI or OEM-proprietary tools.
Designed for Maximum Reliability Passive heatsink design eliminates moving parts and cables.
NVIDIA Parallel DataCache™ Accelerates algorithms such as physics solvers, ray-tracing, and sparse matrix multiplication where data addresses are not known beforehand. This includes a configurable L1 cache per Streaming Multiprocessor block and a unified L2 cache for all of the processor cores.
NVIDIA GigaThread™ Engine Maximizes the throughput by faster context switching that is 10X faster than previous architecture, concurrent kernel execution, and improved thread block scheduling.
Asynchronous Transfer Turbocharges system performance by transferring data over the PCIe bus while the computing cores are crunching other data. Even applications with heavy data-transfer requirements, such as seismic processing, can maximize the computing efficiency by transferring data to local memory before it is needed.
CUDA programming environment with broad support of programming languages and APIs Choose C, C++, OpenCL, DirectCompute, or Fortran to express application parallelism and take advantage of the innovative “Fermi” architecture.
High Speed , PCIe Gen 2.0 Data Transfer Maximizes bandwidth between the host system and the Tesla processors. Enables Tesla systems to work with virtually any PCIe-compliant host system with an open PCIe slot (x8 or x16).
Labels: Miscellaneous , Other mobile phone brands
0 comments:
Post a Comment