AI & Machine Learning Systems Enlist NVIDIA GPUs

Super Micro Computer is showcasing what it believes is the industry's broadest selection of GPU server platforms that support NVIDIA Tesla V100 PCI-E and V100 SXM2 Tensor Core GPU accelerators. For maximum acceleration of highly parallel applications like artificial intelligence (AI), deep learning, self-driving cars, smart cities, health care, big data, high performance computing (HPC), virtual reality and more, Supermicro's 4U system with next-generation NVIDIA NVLink interconnect technology is optimized for maximum performance.


SuperServer 4029GP-TVRT, a part of the NVIDIA HGX-T1 class of GPU-Accelerated Server Platforms, supports eight NVIDIA Tesla V100 32GB SXM2 GPU accelerators with maximum GPU-to-GPU bandwidth for cluster and hyper-scale applications. Incorporating the latest NVIDIA NVLink technology with over five times the bandwidth of PCI-E 3.0, this system features independent GPU and CPU thermal zones to ensure uncompromised performance and stability under the most demanding workloads.

Free Monthly Newsletter

Compelling read? Subscribe to FierceEmbeddedTech!

The embedded tech sector runs the market’s trends. FierceEmbeddedTech subscribers rely on our suite of newsletters as their must-read source for the latest news, developments and analysis impacting their world. Sign up today to get news and updates delivered to your inbox and read on the go.


The GPU systems also support the Tesla P4 that is designed to accelerate inference workloads in any scale-out server.  The hardware accelerated transcode engine in Tesla P4 delivers 35 HD video streams in real-time and allows integrating deep learning into the video transcoding pipeline to enable a new class of smart video applications.


Supermicro also offers the NVIDIA SCX-E3 class of GPU-Accelerated Server Platforms, the performance-optimized 4U SuperServer 4029GR-TRT2 system that can support up to 10 PCI-E NVIDIA Tesla V100 accelerators with Supermicro's innovative and GPU-optimized single-root complex PCI-E design, which dramatically improves GPU peer-to-peer communication performance. For even greater density, the SuperServer 1029GQ-TRT supports up to four NVIDIA Tesla V100 PCI-E GPU accelerators in only 1U of rack space and the new SuperServer 1029GQ-TVRT supports four NVIDIA Tesla V100 SXM2 32GB GPU accelerators in 1U. Both 1029GQ servers are part of the NVIDIA SCX-E2 class of GPU-Accelerated Server Platforms.  


With the convergence of big data analytics and machine learning, the latest NVIDIA GPU architectures, and improved machine learning algorithms, deep learning applications require the processing power of multiple GPUs that must communicate efficiently and effectively to expand the GPU network. Supermicro's single-root GPU system allows multiple NVIDIA GPUs to communicate efficiently to minimize latency and maximize throughput as measured by the NCCL P2PBandwidthTest.


For extended information, checkout Supermicro’s NVIDIA GPU system product lines.

Suggested Articles

Critics are concerned about a false sense of public health safety when temperature scanning is used in hospitals and other settings

Machine learning challenge will look for vocal communication between elephants and other behaviors

Iowa State University researchers are working with NSF grant