NVIDIA announces A100 PCIe accelerator

Published: 22nd Jun 2020, 07:30 GMT   Comments

Today NVIDIA announces a new variant of the A100 Tensor Core accelerator, the A100 PCIe.


NVIDIA announces the availability of its new A100 Ampere-based accelerator with the PCI Express 4.0 interface. The product has the same specifications as the A100 SXM variant except for few details.

The A100 PCIe has a TDP of 250W. According to ComputeBase who quotes NVIDIA, the card does indeed have lower TDP specs. For comparison, the SXM variant has a TDP of 400W. However, despite the lower TDP of the PCIe model, NVIDIA says that the peak power for both models is the same, only during a sustained load will provide 10 to 50% lower performance than SXM4 based variant.

The PCIe model comes with the same 40GB HBM2e memory as the SXM model. It also packs 6912 CUDA cores and 432 Tensor cores, it does however not have RT cores, which are are not needed for the application. The A100 PCIe is an accelerator designed for AI training and interference. It also serves as a compute card for large science applications and large data centers.

The systems with A100 PCIe and A100 SXM accelerators are now available from a broad range of server suppliers.

NVIDIA Compute Accelerator Series (Formely Tesla)
VideoCardz.comA100 PCIeA100 SXMTesla V100sTesla V100Tesla P100
GPU7nm GA1007nm GA10012nm GV10012nm GV10016nm GP100
Die Size
826 mm^2
826 mm^2
815 mm^2
815 mm^2
610 mm^2
54 billion
54 billion
21.1 billion
21.1 billion
15.3 billion
CUDA Cores
Tensor Cores
FP16 Compute
FP32 Compute
FP64 Compute
Boost Clock
~1601 MHz
~1533 MHz
1555 GB/s
1555 GB/s
1134 GB/s
900 GB/s
721 GB/s
Eff. Memory Clock
2430 MHz
2430 MHz
2214 MHz
1760 MHz
1408 MHz
Memory Config.
40GB HBM2e
40GB HBM2e
16GB / 32GB HBM2
Memory Bus
Form FactorPCIe 4.0SXM4PCIe 3.0SXM2 / PCIe 3.0SXM

« press release »

World’s Top System Makers Unveil NVIDIA A100-Powered Servers to Accelerate AI, Data Science and Scientific Computing

ISC Digital—NVIDIA and the world’s leading server manufacturers today announced NVIDIA A100-powered systems in a variety of designs and configurations to tackle the most complex challenges in AI, data science and scientific computing.

More than 50 A100-powered servers from leading vendors around the world — including ASUS, Atos, Cisco, Dell Technologies, Fujitsu, GIGABYTE, Hewlett Packard Enterprise, Inspur, Lenovo, One Stop Systems, Quanta/QCT and Supermicro — are expected following last month’s launch of the NVIDIA Ampere architecture and the NVIDIA A100 GPU.

Availability of the servers varies, with 30 systems expected this summer, and over 20 more by the end of the year.

“Adoption of NVIDIA A100 GPUs into leading server manufacturers’ offerings is outpacing anything we’ve previously seen,” said Ian Buck, vice president and general manager of Accelerated Computing at NVIDIA. “The sheer breadth of NVIDIA A100 servers coming from our partners ensures that customers can choose the very best options to accelerate their data centers for high utilization and low total cost of ownership.”

The first GPU based on the NVIDIA Ampere architecture, the A100 can boost performance by up to 20x over its predecessor — making it the company’s largest leap in GPU performance to date. It features several technical breakthroughs, including a new multi-instance GPU technology enabling a single A100 to be partitioned into as many as seven separate GPUs to handle varying compute jobs; third-generation NVIDIA® NVLink® technology that makes it possible to join several GPUs together to operate as one giant GPU; and new structural sparsity capabilities that can be used to double a GPU’s performance.

NVIDIA also unveiled a PCIe form factor for the A100, complementing the four- and eight-way NVIDIA HGX™ A100 configurations launched last month. The addition of a PCIe version enables server makers to provide customers with a diverse set of offerings — from single A100 GPU systems to servers featuring 10 or more GPUs. These systems accelerate a wide range of compute-intensive workloads, from simulating molecular behavior for drug discovery to building better financial models for mortgage approvals.

Server manufacturers bringing NVIDIA A100-powered systems to their customers include:

  • ASUS will offer the ESC4000A-E10, which can be configured with four A100 PCIe GPUs in a single server.
  • Atos is offering its BullSequana X2415 system with four NVIDIA A100 Tensor Core GPUs.
  • Cisco plans to support NVIDIA A100 Tensor Core GPUs in its Cisco Unified Computing System servers and in its hyperconverged infrastructure system, Cisco HyperFlex.
  • Dell Technologies plans to support NVIDIA A100 Tensor Core GPUs across its PowerEdge servers and solutions that accelerate workloads from edge to core to cloud, just as it supports other NVIDIA GPU accelerators, software and technologies in a wide range of offerings.
  • Fujitsu is bringing A100 GPUs to its PRIMERGY line of servers.
  • GIGABYTE will offer G481-HA0, G492-Z50 and G492-Z51 servers that support up to 10 A100 PCIe GPUs, while the G292-Z40 server supports up to eight.
  • HPE will support A100 PCIe GPUs in the HPE ProLiant DL380 Gen10 Server, and for accelerated HPC and AI workloads, in the HPE Apollo 6500 Gen10 System.
  • Inspur is releasing eight NVIDIA A100-powered systems, including the NF5468M5, NF5468M6 and NF5468A5 using A100 PCIe GPUs, the NF5488M5-D, NF5488A5, NF5488M6 and NF5688M6 using eight-way NVLink, and the NF5888M6 with 16-way NVLink.
  • Lenovo will support A100 PCIe GPUs on select systems, including the Lenovo ThinkSystem SR670 AI-ready server. Lenovo will expand availability across its ThinkSystem and ThinkAgile portfolio in the fall.
  • One Stop Systems will offer its OSS 4UV Gen 4 PCIe expansion system with up to eight NVIDIA A100 PCIe GPUs to allow AI and HPC customers to scale out their Gen 4 servers.
  • Quanta/QCT will offer several QuantaGrid server systems, including D52BV-2U, D43KQ-2U and D52G-4U that support up to eight NVIDIA A100 PCIe GPUs.
  • Supermicro will offer its 4U A+ GPU system, supporting up to eight NVIDIA A100 PCIe GPUs and up to two additional high-performance PCI-E 4.0 expansion slots along with other 1U, 2U and 4U GPU servers.

NVIDIA is expanding its portfolio of NGC-Ready™ certified systems. Working directly with NVIDIA, system vendors can receive NGC-Ready certification for their A100-powered servers. NGC-Ready certification assures customers that systems will deliver the performance required to run AI workloads.

NGC-Ready systems are tested with GPU-optimized AI software from NVIDIA’s NGC™ registry, which is available for NVIDIA GPU-powered systems in data centers, the cloud and at the edge.

NVIDIA A100 Optimized Software Now Available

NVIDIA A100 is supported by NVIDIA Ampere-optimized software, including CUDA 11; new versions of more than 50 CUDA-X™ libraries; NVIDIA Jarvis, a multimodal, conversational AI services framework; NVIDIA Merlin, a deep recommender application framework; the RAPIDS™ suite of open source data science software libraries; and the NVIDIA HPC SDK, which includes compilers, libraries and software tools to maximize developer productivity and the performance and portability of HPC applications.

These powerful software tools enable developers to build and accelerate applications in HPC, genomics, 5G, data science, robotics and more.

« end of the press release »

Comment Policy
  • Comments must be written in English.
  • Comments deemed to be spam or solely promotional in nature will be deleted. Including a link to relevant content is permitted, but comments should be relevant to the post topic.
  • Comments and usernames containing language or concepts that could be deemed offensive will be deleted. Note this may include abusive, threatening, pornographic, offensive, misleading, or libelous language.
  • Comments complaining about the article subject or its source will be removed.
  • A failure to comply with these rules will result in a warning and, in extreme cases, a ban.
  • Please note that comments that attack or harass an individual directly will be deleted and such comments will result in a ban.
  • VideoCardz Moderating Team reserves the right to edit or delete any comments submitted to the site without notice.
  • If you would like to appeal for a comment section ban to be removed, please use this page.
  • If you have any questions about the commenting policy, please let us know through the Contact Page.
  • NEW: Due to the recent increased amount of spam all links will require moderator approval. 
Hide Comment Policy