Find Top SoC Solutions
for AI, Automotive, IoT, Security, Audio & Video...

BFloat16 processing for Neural Networks on Armv8-A

community.arm.com, Aug. 29, 2019 – 

Neural Networks are a key component of Machine Learning (ML) applications. Project Trillium, Arm's heterogeneous ML platform, provides a range of technologies in this field, including instructions that accelerate such applications running on CPUs based on the Arm®v8-A architecture.

The next revision of the Armv8-A architecture will introduce Neon and SVE vector instructions designed to accelerate certain computations using the BFloat16 (BF16) floating-point number format. BF16 has recently emerged as a format tailored specifically to high-performance processing of Neural Networks (NNs). BF16 is a truncated form of the IEEE 754 [ieee754-2008] single-precision representation (IEEE-FP32), which has only 7 fraction bits, instead of 23

Several major CPU and GPU architectures, and Neural Network accelerators (NPUs), have announced an intention to support BF16.

Click here to


Partner with us

Visit our new Partnership Portal for more information.

Submit your material

Submit hot news, product or article.

List your Products

Suppliers, list and add your products for free.

More about D&R Privacy Policy

© 2018 Design And Reuse

All Rights Reserved.

No portion of this site may be copied, retransmitted,
reposted, duplicated or otherwise used without the
express written permission of Design And Reuse.