In this presentation, Banu Nagasundaram from Intel offers an overview of Intel’s Deep Learning Boost technology, featuring integer vector neural network instructions targeting future Intel Xeon scalable processors. These instructions improve throughput of multiply-add operations with int8 and int16 data types and are used to achieve performance gains in low-precision convolution and matrix-matrix multiplication operations used in deep neural networks. Banu walks you through the 8-bit integer convolution implementation made in the Intel MKL-DNN library to demonstrate how this new instruction is used in optimized code.
Learn more: https://www.intel.ai/intel-deep-learning-boost/
and
https://wp.me/p3RLHQ-kcm
Sign up for our insideHPC Newsletter: http://insidehpc.com/newsletter
21. #datacentric
Client: Siemens Healthineers is the
global market leader in diagnostic
imaging. Artificial intelligence (AI) is
transforming care delivery and
expanding precision medicine.
Siemens Healthineers utilizes AI to
help automate and standardize
complex diagnostics to improve
patient outcomes.
Challenge: Segmentation of anatomical regions for medical
imaging requires hardware that delivers increased throughput
for Deep Learning inferencing. This is used to automate and
standardize clinical measurements, such as measurement of
heart chambers, ejection fraction, and strain. Due to increased
data generation and workflow demands, high performance
Deep Learning inference is required to help reduce the burden
on radiologists and cardiologists, enabling improved clinical
decisions in near real-time.
Solution: The next generation Intel® Xeon® Scalable processors
with Intel® Deep Learning Boost, in combination with the Intel®
Distribution of OpenVINO™ toolkit, automate anatomical
measurements in near real-time and improve workflow
efficiency, while maintaining accuracy of the model without the
need for discrete GPU investment.
*Other names and brands may be claimed as the property of others.
Configuration: Based on Siemens Healthineers and Intel Analysis on 2nd Gen Intel® Xeon® Platinum 8280 Processor (28 Cores) with 192GB, DDR4-2933, using Intel® OpenVino™ 2019 R1. HT ON, Turbo ON. CentOS Linux release 7.6.1810, kernel 4.19.5-
1.el7.elrepo.x86_64.Custom topology and dataset (image resolution 288x288). Comparing FP32 vs Int8 w/ Intel® DL Boost performance on the system.
Software and workloads used in performance tests may have been optimized for performance only on Intel microprocessors. Performance tests, such as SYSmark and MobileMark, are measured using specific computer systems, components, software, operations and
functions. Any change to any of those factors may cause the results to vary. You should consult other information and performance tests to assist you in fully evaluating your contemplated purchases, including the performance of that product when combined with other
products. For more complete information visit http://www.intel.com/performance.
Result
5.5xfaster(now201fPS)
This Siemens Healthineers’ feature is currently under development and is not currently for sale.
Using 2nd Generation Intel® Xeon
Scalable with Intel® DL Boost (INT8
precision) compared to FP32
precisionLeft Ventricle, Right Ventricle (short-axis), Left Ventricle, Right Atrium and Left Atrium (long-axis)
27. INTEL® DL BOOST ECOSYSTEM SUPPORT
OPTIMIZED SW &
FRAMEWORKS
SOFTWARE
VENDORS
CLOUD SERVICE
PROVIDERS
ENTERPRISES
VIDEO ANALYSIS
TEXT DETECTION
MEDICAL IMAGE
CLASSIFICATION
FACIAL RECOGNITION
ML INFERENCING