Driving Data Center Performance Through Intel Memory Technology – Intel Chip Chat – Episode 624

 
Share
 

Manage episode 224459019 series 29991
By Discovered by Player FM and our community — copyright is owned by the publisher, not Player FM, and audio streamed directly from their servers.

In this Intel Chip Chat audio podcast with Allyson Klein: Dr. Ziya Ma, vice president of Intel Software and Services Group and director of Data Analytics Technologies, gives Chip Chat listeners a look at data center optimization along with a preview of advancements well underway.\n\nIn their work with the broad industry, Dr. Ma and her team have found that taming the data deluge calls for IT data center managers to unify their big data analytics and AI workflows. As they’ve helped customers overcome the memory constraints involved in data caching, Apache Spark, which supports the convergence of AI on big data, has proven to be a highly effective platform.

Dr. Ma and her team have already provided the community a steady stream of source code contributions and optimizations for Spark. In this interview she reveals that more – and even more exciting work – is underway.

Spark depends on memory to perform and scale. That means optimizing Spark for the revolutionary new Intel Optane DC persistent memory offers performance improvement for the data center.

In one example, Dr. Ma describes benchmark testing where Spark SQL performs eight times faster at a 2.6TB data scale using Intel Optane DC persistent memory than a comparable system using DRAM DIMMs.

With Intel Optane DC persistent memory announced and broadly available in 2019, data centers have the chance to achieve workflow unification along with performance gains and system resilience starting now.

For more information about Intel’s work in this space, go to:
software.intel.com/ai

For more about how Intel is driving advances in the ecosystem, visit:
intel.com/analytics

Share

797 episodes available. A new episode about every day .