Empowering AI Service Delivery with an Edge-to-Cloud Intel Technology Portfolio

Service providers and end users worldwide are seeing the benefits of artificial intelligence (AI) as machine learning algorithms are increasingly used to process the world’s data and enhance our digital services. Using AI to make the most of the data opportunity requires a complete workflow, from data science workstations up to cloud and eventually out to inference devices – not only for processing data, but moving and storing data as well.

Inference is How End Users Interact with AI

Today, we’re moving from training models to deploying them in the real world. Previously, much of the work in AI focused on training – refining the model for the application you hope to render. However, end users for AI services don’t experience training. They experience inference – the rendering of the AI service.

Services often must render results rapidly to be relevant to their end users – whether those are medical professionals, research scientists, or consumers of voice recognition services. As a result, we see more inference in local servers and Internet of Things (IoT) devices at the edge, driven by the need for low-latency, real-time inference results, in addition to the inference on less-time-sensitive data sent the cloud.

Real Value from Data

At Data-Centric Innovation Day, we are excited to highlight several AI deployments delivering rapid, real-world results for a seamless user experience by drawing on the latest additions to Intel’s diverse silicon portfolio, such as 2nd Generation Intel® Xeon® Scalable processors and Intel® Optane™ DC persistent memory.

The Texas Advanced Computing Center (TACC) will use Intel® Xeon® Platinum 8200 processors to power its Frontera system, supporting multi-faceted advanced research for the National Science Foundation. At Data-Centric Innovation Day, we are excited to share that Frontera will also incorporate more than 100 terabytes of Intel Optane DC persistent memory, the first installation of the technology at this scale. This store of persistent memory in close proximity to performant compute will enable simulations, AI algorithms, and in-memory analytics of unprecedented complexity. Frontera will help to reveal what’s possible with massively-parallel AI inference on high-performance computing systems. We eagerly look forward to the discoveries that Frontera will produce.

iFLYTEK is one of the most innovative companies in the People’s Republic of China and supports a variety of voice-based products in industries like communications, music, and intelligent toys. Customers will turn elsewhere if the company can’t process its daily volume of six billion voice recognition transactions expediently. iFLYTEK faces the continual challenge of expanding data center capacity to keep up with increasing customer demand, with total cost of ownership as one of their primary concerns. Adding to this challenge is iFLYTEK’s ongoing expansion into new businesses such as education and medical diagnostics.
For several years, iFLYTEK has actively migrated more of its business to Intel architecture, including 2nd Gen Intel Xeon Scalable processors with Intel® Deep Learning Boost (Intel DL Boost). The AI giant’s reliance on Intel reiterates the capabilities of Intel solutions to deliver leading AI products in a cost-effective manner to end users.

A Comprehensive AI Portfolio

Data-Centric Innovation Day features the debuts of technologies that will be under the hood of systems running complex AI workloads alongside the traditional data center and cloud applications at which Intel Xeon Scalable processor based systems excel.

  • With the addition of Intel DL Boost – essentially an AI inference accelerator built into the CPU – 2nd Generation Intel Xeon Scalable processorshave demonstrated AI inference throughput increases of 14X in comparison to Intel Xeon Scalable processors at their launch in July 2017. [1]
  • Intel Optane DC persistent memory fundamentally re-architects the storage pyramid, bringing large amounts of persistent memory nearer to compute than ever before and facilitating current AI use cases like high-content image analysis and allowing memory bottlenecks for the most complex current and future deep learning applications to be broken.
  • The Intel® Optane™ DC SSD P4800X series of solid state drives offer reduced data access latency for drives under write pressure, enabling rapid data analysis even as more raw data and results are stored and increasing data availability for mission-critical applications.
  • Intel® Select Solutions for AI Inferencing take the guesswork out of AI hardware selection and deployment, assembling all of the components needed to deliver an enterprise-ready inferencing system. With 2nd Gen Intel Xeon Scalable processors with Intel DL Boost, Intel Optane SSDs, and the Intel Distribution of OpenVINO Toolkit, these solutions maximize the capabilities of Intel architecture for AI inference, all within a package that makes it easier for customers to select the right system to suit their inference needs.

The results customers are realizing with versatile, efficient, performant Intel architecture, especially when dealing with very large workloads, demonstrate again and again that real-world AI solutions require systems able to balance the need to move, process, and store larger and larger quantities of data. It’s not only about whether you have the right processors, accelerators, and storage. It’s how you balance the entire system between compute, acceleration, memory, memory access, and interconnect.

AI Results at the Speeds End Users Demand

Ultimately, AI service providers will succeed or fail based on the quality of experience provided to their own end customers. In use cases like those discussed here, the speed of AI inference and accuracy of results delivered will determine whether a product is relevant or irrelevant to end users. The products Intel announced at Data-Centric Innovation Day will enable complete AI systems for scalable, deployable, real-world results at the speeds end users require. For more on Intel’s product portfolio for AI, please visit www.intel.ai.