Credit: Dell Technologies Artificial intelligence (AI) comes to our minds whenever organizations are seeking new, innovative approaches to enhance and expand their businesses. Typical attributes for such workloads are “use case centric”, “data scientist specific”, and “innovative”. Some examples include autonomous farming, self-driving cars, or interactive voice dialog systems. On the other hand, high performance computing (HPC) is often seen as highly specialized and expansive, serving a wide range of custom-written applications by research engineers, or some of the major industry HPC software stacks like fluid dynamics modelling or crash simulations. For many years now, both disciplines have been treated separately, developing their own ecosystem of specialized hardware, software stacks, and operational models. But if you take a look from the outside, neither workload is far away from the other regarding their basic requirements. Luckily, two recent developments in technology have made it possible for organizations to standardize on a common production environment: Common AI frameworks like TensorFlow have become more mature and established Server Infrastructure and Networking technology advantages deliver greater performance Given those two developments, a new and broader definition of HPC is possible which includes the following four areas: Traditional HPC – like weather forecasting and oil exploration Data Centric HPC – like financial modelling or genomics High Performance Data Analytics – like fraud detection or personalized medicine Artificial Intelligence – including deep and machine learning Now that more organizations are running ─ or is planning to run ─ AI initiatives, almost every organization is now a home for HPC. However, various studies, including one from 451 Research, have revealed several issues that interfere with the success of those initiatives. I want to point out two of the most relevant ones. First, the lack of expertise and getting enough skilled resources is a challenge for nearly every organization. Thus, having skilled workers is crucial to working as efficiently as possible. Secondly, limited budgets within IT organizations restricts the number of choices available for solving a given problem. While there is always a best-of-breed solution available from a technology point of view, the cost structure, complexity and reusability of such a solution is far from optimal from a price/performance perspective, so a “good-enough” model approach yields to a much more flexible, reusable solution – typically with much lower CAPEX and OPEX costs. This all highlights the need to have a standardized HPC environment that ideally fits directly into the existing IT infrastructure. To help organizations with this, Dell Technologies has created a set of easy to consume, workload orientated architectures as Ready Solutions or Ready Architectures. The advantage of such an approach is that it gives flexibility of choices but builds on standards. The recipe is not too complicated: Standard x86 Servers with latest Intel CPU, and optional accelerator cards (only if required) High-speed Ethernet Networking (25/100 Gbit) (no need for Fibre Channel or Infiniband in most cases) A centralized data lake to serve data to the compute nodes. (i.e. PowerScale – additional parallel Filesystems are only required for certain use cases) Automation software that delivers docker and Kubernetes, workload scheduling, and can provide a supported stack of AI tools and frameworks. (i.e. Bright Cluster Manager) Services and Support to support the whole system from a single point of view. A data scientist centric working environment (i.e. Jupyter notebooks) With such a system, most of the HPC workloads described above can be run in a very efficient way in regards to IT operations, cost structure and user acceptance. In conclusion: Only specialize in IT infrastructure when you really need it. Most of the demands can be satisfied with well-designed standard hardware components, an optimized software stack and automation capabilities. So, from this angle, AI is indeed just another HPC workload. Related content brandpost The steep cost of a poor data management strategy Without a data management strategy, organizations stall digital progress, often putting their business trajectory at risk. Here’s how to move forward. By Jay Limbasiya, Global AI, Analytics, & Data Management Business Development, Unstructured Data Solutions, Dell Technologies Jun 09, 2023 6 mins Data Management brandpost Democratizing HPC with multicloud to accelerate engineering innovations Cloud for HPC is facilitating broader access to high performance computing and accelerating innovations and opportunities for all types of organizations. By Tanya O'Hara Jun 01, 2023 6 mins Multi Cloud brandpost Solving 3 key IT challenges to unlock business innovation Dell and Microsoft are integrating strengths to help organizations unlock innovation with cloud-like agility across on-premises, edge, and cloud environments. By Vikram Belapurkar, Product Marketing, Multicloud, and Software-defined Infrastructure Platforms, Dell Technologies May 23, 2023 4 mins Hybrid Cloud brandpost How to Make the Quantum (Computing) Leap Three steps to start deploying quantum computing applications. By Mike Robillard, Senior Distinguished Engineer, Office of the CTO, Dell Technologies and Victor Fong, Distinguished Engineer, Office of the CTO, Dell Technologies May 08, 2023 7 mins Digital Transformation Podcasts Videos Resources Events SUBSCRIBE TO OUR NEWSLETTER From our editors straight to your inbox Get started by entering your email address below. Please enter a valid email address Subscribe