Delivering AI and HPC workloads at the edge has historically been a challenge. Form factor, latency, and power can all lead to key limitations on the edge. For this discussion, the edge means any compute workloads taking place outside of both cloud and traditional on-prem data centers.
Recently, however, key advancements in technology will allow higher performance at the edge. Powerful new technologies, including NVIDIA GPU, InfiniBand, and Ethernet, can deliver the performance required for AI and HPC at the edge, while still allowing strong ROI.
In addition, 5G networking will both drive the explosion of new connected devices that generate more valuable real-time data and improve the abilities of AI deep learning (DL) systems. Accomplishing DL in the field without latency or bandwidth issues can become a reality with the advancement of 5G.
Strong design teams can create reference architectures that consider both the accelerated computing needs and edge form factors involved in these types of workloads. A reference architecture is a starting point for customized design, with critical elements already thought through and tested.
Each reference architecture addresses engineering, testing, and optimization for power, latency, and related concerns of resource-hungry applications. They also consider space, size, ruggedization, and other unique issues facing in-field deployments.
Here are several examples of reference architecture designs that allow you to optimize infrastructure to accommodate high-performance edge workloads.
This first cluster is a small-footprint, air-gapped cloud environment-in-a-box designed for performance, reliability, portability, and security at the edge. This solution provides a high-performance computing environment to support critical operations at the edge, such as secure development environments, remote location computing, and more.
The design is a small rack unit configuration (about 6U) to ensure the solution is portable. The cluster makes expansion simple, allowing you to add compute and storage resources and even scale up to a full rack if you choose. It also allows pooled resilient storage and features a configurable network speed to meet your needs.
This solution supports numerous users in remote, air-gapped areas offering all the performance they need to deploy workloads at the edge without latency going back to the cloud. It’s ideal for operating environments that simply can’t be connected to the cloud, featuring dramatically simplified development and management without sacrificing performance or security.
Ideal Use Cases
An edge appliance allows you to cost-effectively access the storage and compute power you need via a local resource ruggedized for the edge, capable of running complex operations in harsh, non-data center environments. This appliance uses standard hardware but provides a complete software solution stack already bundled together to meet the needs of the deployment.
The key differentiator for these edge appliances is the ruggedized chassis, built to MIL-SPEC attributes, able to run in any difficult operating environments.
Ideal Use Cases
This cluster can process AI inference workloads at the edge in real time while protecting equipment from environmental hazards. It operates on limited power, in small footprints, in broad temperature ranges, and it resists dust and moisture.
Ideal Use Cases
This cluster is purpose-built to support cloud and accelerated workloads at the edge without the need for virtualization. This allows you to get bare metal performance of HPC, AI, and even ML at the edge.
It’s also very flexible because composable disaggregated infrastructure (CDI) allows you to reconfigure systems as needs dictate. Resources such as GPUs, FPGAs, NVMe storage, and more, are connected via PCIe-connected resources so you can scale each element independently. Your deployment can be reconfigured based on the workload without losing performance because of the flexibility with your hardware. If you’re in the field and looking to deploy several different workloads, CDI will allow you to dynamically reconfigure the hardware so that each workload has a very specific optimized solution.
Ideal Use Cases
Silicon Mechanics is an engineering firm providing custom, best-in-class solutions for HPC/AI, storage, and networking, based on open standards. The experts at Silicon Mechanics understand that implementing HPC and AI on the edge requires a strong understanding of both computing and form factor. That’s why we created a series of reference architectures for specific types of edge deployments and workloads.
Get a more comprehensive understanding of Silicon Mechanics edge reference architectures and what they can do for your organization at www.siliconmechanics.com/edge.
Silicon Mechanics, Inc. is one of the world’s largest private providers of high-performance computing (HPC), artificial intelligence (AI), and enterprise storage solutions. Since 2001, Silicon Mechanics’ clients have relied on its custom-tailored open-source systems and professional services expertise to overcome the world’s most complex computing challenges. With thousands of clients across the aerospace and defense, education/research, financial services, government, life sciences/healthcare, and oil and gas sectors, Silicon Mechanics solutions always come with “Expert Included” SM. Learn more at www.siliconmechanics.com.
AI workloads have led to new hardware standards that need GPU acceleration. GPU-accelerated neural net training/inference is now an advantageous option.READ MORE
Achieve the advantages of the InterPlanetary File System (IPFS) by building the ideal infrastructure for IPFS deployments and workloads.READ MORE
Our engineers are not only experts in traditional HPC and AI technologies, we also routinely build complex rack-scale solutions with today's newest innovations so that we can design and build the best solution for your unique needs.
Talk to an engineer and see how we can help solve your computing challenges today.