Artificial intelligence (AI) comes to our minds whenever organizations are seeking new, innovative approaches to enhance and expand their businesses. Typical attributes for such workloads are \u201cuse case centric\u201d, \u201cdata scientist specific\u201d, and \u201cinnovative\u201d. Some examples include autonomous farming, self-driving cars, or interactive voice dialog systems.\nOn the other hand, high performance computing (HPC) is often seen as highly specialized and expansive, serving a wide range of custom-written applications by research engineers, or some of the major industry HPC software stacks like fluid dynamics modelling or crash simulations.\nFor many years now, both disciplines have been treated separately, developing their own ecosystem of specialized hardware, software stacks, and operational models. But if you take a look from the outside, neither workload is far away from the other regarding their basic requirements. Luckily, two recent developments in technology have made it possible for organizations to standardize on a common production environment:\n\nCommon AI frameworks like TensorFlow have become more mature and established\nServer Infrastructure and Networking technology advantages deliver greater performance\n\nGiven those two developments, a new and broader definition of HPC is possible which includes the following four areas:\n\nTraditional HPC \u2013 like weather forecasting and oil exploration\nData Centric HPC \u2013 like financial modelling or genomics\nHigh Performance Data Analytics \u2013 like fraud detection or personalized medicine\nArtificial Intelligence \u2013 including deep and machine learning\n\nNow that more organizations are running \u2500 or is planning to run \u2500 AI initiatives, almost every organization is now a home for HPC. However, various studies, including one from 451 Research, have revealed several issues that interfere with the success of those initiatives. I want to point out two of the most relevant ones.\nFirst, the lack of expertise and getting enough skilled resources is a challenge for nearly every organization. Thus, having skilled workers is crucial to working as efficiently as possible. Secondly, limited budgets within IT organizations restricts the number of choices available for solving a given problem.\nWhile there is always a best-of-breed solution available from a technology point of view, the cost structure, complexity and reusability of such a solution is far from optimal from a price\/performance perspective, so a \u201cgood-enough\u201d model approach yields to a much more flexible, reusable solution \u2013 typically with much lower CAPEX and OPEX costs.\nThis all highlights the need to have a standardized HPC environment that ideally fits directly into the existing IT infrastructure.\nTo help organizations with this, Dell Technologies has created a set of easy to consume, workload orientated architectures as Ready Solutions or Ready Architectures. The advantage of such an approach is that it gives flexibility of choices but builds on standards.\nThe recipe is not too complicated:\n\nStandard x86 Servers with latest Intel CPU, and optional accelerator cards (only if required)\nHigh-speed Ethernet Networking (25\/100 Gbit) (no need for Fibre Channel or Infiniband in most cases)\nA centralized data lake to serve data to the compute nodes. (i.e. PowerScale - additional parallel Filesystems are only required for certain use cases)\nAutomation software that delivers docker and Kubernetes, workload scheduling, and can provide a supported stack of AI tools and frameworks. (i.e. Bright Cluster Manager)\nServices and Support to support the whole system from a single point of view.\nA data scientist centric working environment (i.e. Jupyter notebooks)\n\nWith such a system, most of the HPC workloads described above can be run in a very efficient way in regards to IT operations, cost structure and user acceptance.\nIn conclusion: Only specialize in IT infrastructure when you really need it. Most of the demands can be satisfied with well-designed standard hardware components, an optimized software stack and automation capabilities. So, from this angle, AI is indeed just another HPC workload.