Artwork

Content provided by Intel Corporation. All podcast content including episodes, graphics, and podcast descriptions are uploaded and provided directly by Intel Corporation or their podcast platform partner. If you believe someone is using your copyrighted work without your permission, you can follow the process outlined here https://player.fm/legal.
Player FM - Podcast App
Go offline with the Player FM app!

Accelerating AI Inference with Intel® Deep Learning Boost – Intel® Chip Chat episode 632

11:47
 
Share
 

Manage episode 226984257 series 1210447
Content provided by Intel Corporation. All podcast content including episodes, graphics, and podcast descriptions are uploaded and provided directly by Intel Corporation or their podcast platform partner. If you believe someone is using your copyrighted work without your permission, you can follow the process outlined here https://player.fm/legal.
When Intel previewed an array of data-centric innovations in August 2018, one that captured media attention was Intel® Deep Learning Boost, an embedded AI accelerator in the CPU designed to speed deep learning inference workloads. Intel DL Boost will make its initial appearance in the upcoming generation of Intel® Xeon® Scalable processors code-named Cascade Lake. In this Chip Chat podcast, Intel Data-centric Platform Marketing Director Jason Kennedy shares details about the optimization behind some impressive test results. The key to Intel DL Boost – and its performance kick – is augmentation of the existing Intel® Advanced Vector Extensions 512 (Intel® AVX-512) instruction set. This innovation significantly accelerates inference performance for deep learning workloads optimized to use vector neural network instructions (VNNI). Image classification, language translation, object detection, and speech recognition are just a few examples of workloads that can benefit. Early tests have shown image recognition 11 times faster using a similar configuration than with current-generation Intel Xeon Scalable processors when launched in July 2017. Current projections estimate 17 times faster inference throughput benefit with Intel® Optimized Caffe ResNet-50 and Intel Deep Learning Boost that can be achieved with a new class of advance performance CPUs debuting in the upcoming generation. For more information about AI activities across Intel visit ai.intel.com. Software and workloads used in performance tests may have been optimized for performance only on Intel microprocessors. Performance tests, such as SYSmark and MobileMark, are measured using specific computer systems, components, software, operations and functions. Any change to any of those factors may cause the results to vary. You should consult other information and performance tests to assist you in fully evaluating your contemplated purchases, including the performance of that product when combined with other products. For more information go to www.intel.com/benchmarks. Performance results are based on testing or projections as of 6/2017 to 11/7/2018 and may not reflect all publicly available security updates. See configuration disclosures in https://intel.ly/2Mw7KKE and https://intel.ly/2VUvY2I for details. No product can be absolutely secure. [1] Future Intel Xeon Scalable processors codenamed Cascade Lake provide up to 11x inference performance with Intel® Optimized Caffe, ResNet-50, and Intel® Deep Learning Boost (VNNI) in comparison to current generation Intel Xeon Scalable Platinum 8180 Processor with FP32 instructions (at launch, measured July, 2017), for details see https://intel.ly/2Mw7KKE, slide 33. [2] Future Intel Xeon Scalable processors codenamed Cascade Lake advanced performance provide up to 17x images per second (deep learning inference) in comparison to current generation Intel Xeon Scalable Platinum 8180 Processor (at launch, measured July, 2017), for details see https://intel.ly/2VUvY2I.
  continue reading

172 episodes

Artwork
iconShare
 
Manage episode 226984257 series 1210447
Content provided by Intel Corporation. All podcast content including episodes, graphics, and podcast descriptions are uploaded and provided directly by Intel Corporation or their podcast platform partner. If you believe someone is using your copyrighted work without your permission, you can follow the process outlined here https://player.fm/legal.
When Intel previewed an array of data-centric innovations in August 2018, one that captured media attention was Intel® Deep Learning Boost, an embedded AI accelerator in the CPU designed to speed deep learning inference workloads. Intel DL Boost will make its initial appearance in the upcoming generation of Intel® Xeon® Scalable processors code-named Cascade Lake. In this Chip Chat podcast, Intel Data-centric Platform Marketing Director Jason Kennedy shares details about the optimization behind some impressive test results. The key to Intel DL Boost – and its performance kick – is augmentation of the existing Intel® Advanced Vector Extensions 512 (Intel® AVX-512) instruction set. This innovation significantly accelerates inference performance for deep learning workloads optimized to use vector neural network instructions (VNNI). Image classification, language translation, object detection, and speech recognition are just a few examples of workloads that can benefit. Early tests have shown image recognition 11 times faster using a similar configuration than with current-generation Intel Xeon Scalable processors when launched in July 2017. Current projections estimate 17 times faster inference throughput benefit with Intel® Optimized Caffe ResNet-50 and Intel Deep Learning Boost that can be achieved with a new class of advance performance CPUs debuting in the upcoming generation. For more information about AI activities across Intel visit ai.intel.com. Software and workloads used in performance tests may have been optimized for performance only on Intel microprocessors. Performance tests, such as SYSmark and MobileMark, are measured using specific computer systems, components, software, operations and functions. Any change to any of those factors may cause the results to vary. You should consult other information and performance tests to assist you in fully evaluating your contemplated purchases, including the performance of that product when combined with other products. For more information go to www.intel.com/benchmarks. Performance results are based on testing or projections as of 6/2017 to 11/7/2018 and may not reflect all publicly available security updates. See configuration disclosures in https://intel.ly/2Mw7KKE and https://intel.ly/2VUvY2I for details. No product can be absolutely secure. [1] Future Intel Xeon Scalable processors codenamed Cascade Lake provide up to 11x inference performance with Intel® Optimized Caffe, ResNet-50, and Intel® Deep Learning Boost (VNNI) in comparison to current generation Intel Xeon Scalable Platinum 8180 Processor with FP32 instructions (at launch, measured July, 2017), for details see https://intel.ly/2Mw7KKE, slide 33. [2] Future Intel Xeon Scalable processors codenamed Cascade Lake advanced performance provide up to 17x images per second (deep learning inference) in comparison to current generation Intel Xeon Scalable Platinum 8180 Processor (at launch, measured July, 2017), for details see https://intel.ly/2VUvY2I.
  continue reading

172 episodes

All episodes

×
 
Loading …

Welcome to Player FM!

Player FM is scanning the web for high-quality podcasts for you to enjoy right now. It's the best podcast app and works on Android, iPhone, and the web. Signup to sync subscriptions across devices.

 

Quick Reference Guide